var/home/core/zuul-output/0000755000175000017500000000000015111431576014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111436232015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004022425315111436223017675 0ustar rootrootNov 25 22:56:43 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 22:56:43 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:43 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 22:56:44 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 22:56:44 crc kubenswrapper[4761]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.722478 4761 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728771 4761 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728800 4761 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728810 4761 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728819 4761 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728828 4761 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728835 4761 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728843 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728852 4761 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728860 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728869 4761 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728877 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728884 4761 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728892 4761 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728899 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728908 4761 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728919 4761 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728930 4761 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728939 4761 feature_gate.go:330] unrecognized feature gate: Example Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728948 4761 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728957 4761 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728965 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728973 4761 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728984 4761 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.728995 4761 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729005 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729013 4761 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729021 4761 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729029 4761 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729038 4761 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729056 4761 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729065 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729073 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729081 4761 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729089 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729096 4761 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729104 4761 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729114 4761 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729121 4761 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729129 4761 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729139 4761 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729149 4761 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729159 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729195 4761 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729205 4761 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729214 4761 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729223 4761 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729230 4761 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729239 4761 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729248 4761 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729256 4761 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729264 4761 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729273 4761 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729285 4761 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729294 4761 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729302 4761 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729310 4761 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729318 4761 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729326 4761 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729334 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729342 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729350 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729359 4761 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729367 4761 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729375 4761 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729382 4761 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729391 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729399 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729407 4761 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729414 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729422 4761 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.729430 4761 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730371 4761 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730394 4761 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730410 4761 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730422 4761 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730434 4761 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730444 4761 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730466 4761 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730477 4761 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730486 4761 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730495 4761 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730507 4761 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730517 4761 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730526 4761 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730535 4761 flags.go:64] FLAG: --cgroup-root="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730544 4761 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730553 4761 flags.go:64] FLAG: --client-ca-file="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730562 4761 flags.go:64] FLAG: --cloud-config="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730571 4761 flags.go:64] FLAG: --cloud-provider="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730579 4761 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730590 4761 flags.go:64] FLAG: --cluster-domain="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730599 4761 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730609 4761 flags.go:64] FLAG: --config-dir="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730618 4761 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730627 4761 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730638 4761 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730648 4761 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730657 4761 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730667 4761 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730676 4761 flags.go:64] FLAG: --contention-profiling="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730685 4761 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730720 4761 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730730 4761 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730739 4761 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730750 4761 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730760 4761 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730769 4761 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730778 4761 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730787 4761 flags.go:64] FLAG: --enable-server="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730796 4761 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730809 4761 flags.go:64] FLAG: --event-burst="100" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730818 4761 flags.go:64] FLAG: --event-qps="50" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730827 4761 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730838 4761 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730847 4761 flags.go:64] FLAG: --eviction-hard="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730858 4761 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730867 4761 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730876 4761 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730885 4761 flags.go:64] FLAG: --eviction-soft="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730894 4761 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730903 4761 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730912 4761 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730921 4761 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730930 4761 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730941 4761 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730950 4761 flags.go:64] FLAG: --feature-gates="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730961 4761 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730970 4761 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730979 4761 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730989 4761 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.730998 4761 flags.go:64] FLAG: --healthz-port="10248" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731007 4761 flags.go:64] FLAG: --help="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731016 4761 flags.go:64] FLAG: --hostname-override="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731024 4761 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731034 4761 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731047 4761 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731059 4761 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731073 4761 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731084 4761 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731096 4761 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731107 4761 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731118 4761 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731130 4761 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731142 4761 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731154 4761 flags.go:64] FLAG: --kube-reserved="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731166 4761 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731177 4761 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731190 4761 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731201 4761 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731213 4761 flags.go:64] FLAG: --lock-file="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731224 4761 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731236 4761 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731246 4761 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731259 4761 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731269 4761 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731279 4761 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731288 4761 flags.go:64] FLAG: --logging-format="text" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731297 4761 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731307 4761 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731316 4761 flags.go:64] FLAG: --manifest-url="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731325 4761 flags.go:64] FLAG: --manifest-url-header="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731337 4761 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731346 4761 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731357 4761 flags.go:64] FLAG: --max-pods="110" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731366 4761 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731375 4761 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731384 4761 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731393 4761 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731402 4761 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731411 4761 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731420 4761 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731442 4761 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731451 4761 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731461 4761 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731470 4761 flags.go:64] FLAG: --pod-cidr="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731479 4761 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731494 4761 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731503 4761 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731513 4761 flags.go:64] FLAG: --pods-per-core="0" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731522 4761 flags.go:64] FLAG: --port="10250" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731531 4761 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731540 4761 flags.go:64] FLAG: --provider-id="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731549 4761 flags.go:64] FLAG: --qos-reserved="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731557 4761 flags.go:64] FLAG: --read-only-port="10255" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731567 4761 flags.go:64] FLAG: --register-node="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731575 4761 flags.go:64] FLAG: --register-schedulable="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731585 4761 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731599 4761 flags.go:64] FLAG: --registry-burst="10" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731608 4761 flags.go:64] FLAG: --registry-qps="5" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731617 4761 flags.go:64] FLAG: --reserved-cpus="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731626 4761 flags.go:64] FLAG: --reserved-memory="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731637 4761 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731646 4761 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731655 4761 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731664 4761 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731673 4761 flags.go:64] FLAG: --runonce="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731682 4761 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731691 4761 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731729 4761 flags.go:64] FLAG: --seccomp-default="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731738 4761 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731747 4761 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731756 4761 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731766 4761 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731775 4761 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731784 4761 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731792 4761 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731801 4761 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731812 4761 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731824 4761 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731836 4761 flags.go:64] FLAG: --system-cgroups="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731849 4761 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731867 4761 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731879 4761 flags.go:64] FLAG: --tls-cert-file="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731890 4761 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731904 4761 flags.go:64] FLAG: --tls-min-version="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731915 4761 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731924 4761 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731934 4761 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731943 4761 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731952 4761 flags.go:64] FLAG: --v="2" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731964 4761 flags.go:64] FLAG: --version="false" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731975 4761 flags.go:64] FLAG: --vmodule="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731986 4761 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.731995 4761 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732212 4761 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732222 4761 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732232 4761 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732246 4761 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732255 4761 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732263 4761 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732270 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732278 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732288 4761 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732299 4761 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732309 4761 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732323 4761 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732335 4761 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732346 4761 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732359 4761 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732373 4761 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732384 4761 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732395 4761 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732404 4761 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732414 4761 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732423 4761 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732435 4761 feature_gate.go:330] unrecognized feature gate: Example Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732445 4761 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732455 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732464 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732475 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732488 4761 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732498 4761 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732507 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732518 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732528 4761 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732538 4761 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732547 4761 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732556 4761 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732566 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732579 4761 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732587 4761 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732595 4761 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732605 4761 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732614 4761 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732623 4761 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732630 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732638 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732646 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732654 4761 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732662 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732669 4761 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732677 4761 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732685 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732693 4761 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732753 4761 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732761 4761 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732769 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732776 4761 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732785 4761 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732793 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732801 4761 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732811 4761 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732819 4761 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732827 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732838 4761 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732848 4761 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732856 4761 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732865 4761 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732873 4761 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732881 4761 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732890 4761 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732903 4761 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732911 4761 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732918 4761 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.732926 4761 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.733899 4761 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.747927 4761 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.748266 4761 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748419 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748440 4761 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748449 4761 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748458 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748472 4761 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748483 4761 feature_gate.go:330] unrecognized feature gate: Example Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748492 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748501 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748509 4761 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748518 4761 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748528 4761 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748537 4761 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748546 4761 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748555 4761 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748563 4761 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748571 4761 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748579 4761 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748587 4761 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748595 4761 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748604 4761 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748614 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748622 4761 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748629 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748637 4761 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748645 4761 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748653 4761 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748661 4761 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748669 4761 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748676 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748686 4761 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748769 4761 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748783 4761 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748791 4761 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748799 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748807 4761 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748816 4761 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748823 4761 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748832 4761 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748841 4761 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748849 4761 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748857 4761 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748867 4761 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748877 4761 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748886 4761 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748894 4761 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748902 4761 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748910 4761 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748918 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748926 4761 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748935 4761 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748943 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748952 4761 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748960 4761 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748968 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748976 4761 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748987 4761 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.748997 4761 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749006 4761 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749014 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749022 4761 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749030 4761 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749039 4761 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749046 4761 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749054 4761 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749062 4761 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749070 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749079 4761 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749086 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749094 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749102 4761 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749110 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.749125 4761 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749369 4761 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749386 4761 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749395 4761 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749404 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749412 4761 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749419 4761 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749428 4761 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749436 4761 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749443 4761 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749452 4761 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749460 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749468 4761 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749476 4761 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749484 4761 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749491 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749501 4761 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749509 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749517 4761 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749526 4761 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749534 4761 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749542 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749550 4761 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749557 4761 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749565 4761 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749573 4761 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749581 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749589 4761 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749596 4761 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749607 4761 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749618 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749627 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749637 4761 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749644 4761 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749654 4761 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749663 4761 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749672 4761 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749680 4761 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749691 4761 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749730 4761 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749743 4761 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749752 4761 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749761 4761 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749771 4761 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749780 4761 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749788 4761 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749796 4761 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749805 4761 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749814 4761 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749823 4761 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749831 4761 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749839 4761 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749848 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749855 4761 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749863 4761 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749871 4761 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749878 4761 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749886 4761 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749894 4761 feature_gate.go:330] unrecognized feature gate: Example Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749902 4761 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749910 4761 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749918 4761 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749926 4761 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749937 4761 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749946 4761 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749954 4761 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749962 4761 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749972 4761 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749983 4761 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.749992 4761 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.750001 4761 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.750009 4761 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.750021 4761 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.750340 4761 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.756633 4761 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.756850 4761 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.759091 4761 server.go:997] "Starting client certificate rotation" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.759145 4761 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.759383 4761 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-14 00:37:32.285458928 +0000 UTC Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.759526 4761 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1177h40m47.525936861s for next certificate rotation Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.795916 4761 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.799658 4761 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.816955 4761 log.go:25] "Validated CRI v1 runtime API" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.857019 4761 log.go:25] "Validated CRI v1 image API" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.859984 4761 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.869622 4761 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-22-52-56-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.869683 4761 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.896288 4761 manager.go:217] Machine: {Timestamp:2025-11-25 22:56:44.893569054 +0000 UTC m=+0.626914909 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:262ec32c-2342-4d0c-b69e-19ea6b9dc450 BootID:a69c5537-0b60-4061-bda8-fe831ae1bd6e Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:17:cd:8e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:17:cd:8e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5e:2a:c7 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3f:03:54 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:dc:b6:83 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b2:88:5a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0e:f9:ca:da:15:77 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:b6:93:43:9b:0c:37 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.896614 4761 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.896886 4761 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.898093 4761 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.898333 4761 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.898383 4761 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.899041 4761 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.899096 4761 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.899644 4761 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.899727 4761 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.900004 4761 state_mem.go:36] "Initialized new in-memory state store" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.900213 4761 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.907325 4761 kubelet.go:418] "Attempting to sync node with API server" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.907404 4761 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.907448 4761 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.907523 4761 kubelet.go:324] "Adding apiserver pod source" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.907556 4761 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.912680 4761 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.914075 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.914141 4761 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.914223 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.914290 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.914317 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.916913 4761 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918506 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918545 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918560 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918573 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918595 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918609 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918623 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918647 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918663 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918677 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918729 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.918743 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.919594 4761 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.920888 4761 server.go:1280] "Started kubelet" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.921502 4761 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.921555 4761 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.921627 4761 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.922231 4761 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 22:56:44 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.923907 4761 server.go:460] "Adding debug handlers to kubelet server" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.924931 4761 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.924975 4761 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.925178 4761 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.925197 4761 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.925423 4761 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.925347 4761 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 16:33:50.288428455 +0000 UTC Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.925548 4761 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 809h37m5.362886403s for next certificate rotation Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.926788 4761 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.927294 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="200ms" Nov 25 22:56:44 crc kubenswrapper[4761]: W1125 22:56:44.927902 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.928012 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.929958 4761 factory.go:55] Registering systemd factory Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.929999 4761 factory.go:221] Registration of the systemd container factory successfully Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.930851 4761 factory.go:153] Registering CRI-O factory Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.936441 4761 factory.go:221] Registration of the crio container factory successfully Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.936648 4761 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.936776 4761 factory.go:103] Registering Raw factory Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.936819 4761 manager.go:1196] Started watching for new ooms in manager Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.938242 4761 manager.go:319] Starting recovery of all containers Nov 25 22:56:44 crc kubenswrapper[4761]: E1125 22:56:44.934277 4761 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b61f9e27d99d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 22:56:44.920314328 +0000 UTC m=+0.653660193,LastTimestamp:2025-11-25 22:56:44.920314328 +0000 UTC m=+0.653660193,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949288 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949367 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949391 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949413 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949432 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949450 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949471 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949489 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949511 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949530 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949548 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949566 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949585 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949609 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949631 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949648 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949666 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.949685 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.951874 4761 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.951928 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952083 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952118 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952149 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952182 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952211 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952239 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952267 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952298 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952329 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952359 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952389 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952415 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952441 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952496 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952579 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952613 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952688 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952821 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952853 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952886 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952915 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952945 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.952974 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953003 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953029 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953055 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953083 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953109 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953135 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953163 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953195 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953223 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953251 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953331 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953367 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953400 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953428 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953455 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953481 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953507 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953532 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953557 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953582 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953607 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953634 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953660 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953684 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953748 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953779 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953805 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953837 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953865 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953892 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953918 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953946 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.953976 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954004 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954033 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954060 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954087 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954117 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954148 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954173 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954200 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954225 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954251 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954278 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954309 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954342 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954372 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954402 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954430 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954455 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954484 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954510 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954536 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954561 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954587 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954613 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954639 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954664 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954687 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954749 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954784 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954809 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954845 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954876 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954905 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954934 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954966 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.954992 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955019 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955045 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955073 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955100 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955129 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955159 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955188 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955216 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955248 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955276 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955305 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955335 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955360 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955387 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955413 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955442 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955470 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955497 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955525 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955554 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955580 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955606 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955632 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955658 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955687 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955775 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955802 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955831 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955862 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955891 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955921 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955951 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.955983 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956012 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956100 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956133 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956166 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956194 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956224 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956253 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956284 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956312 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956339 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956364 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956392 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956420 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956452 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956481 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956509 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956539 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956566 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956595 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956623 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956655 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956685 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956782 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956821 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956849 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956878 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956904 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956935 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956962 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.956989 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957017 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957046 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957076 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957109 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957138 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957165 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957193 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957221 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957247 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957282 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957309 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957341 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957371 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957400 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957429 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957457 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957485 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957512 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957543 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957573 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957604 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957633 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957662 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957691 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957759 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957796 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957822 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957848 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957876 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957906 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957934 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957961 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.957989 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.958015 4761 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.958042 4761 reconstruct.go:97] "Volume reconstruction finished" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.958060 4761 reconciler.go:26] "Reconciler: start to sync state" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.972138 4761 manager.go:324] Recovery completed Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.990894 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.993682 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.993788 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.993807 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.995141 4761 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.995188 4761 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 22:56:44 crc kubenswrapper[4761]: I1125 22:56:44.995232 4761 state_mem.go:36] "Initialized new in-memory state store" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.005777 4761 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.008829 4761 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.009819 4761 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.009933 4761 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.010107 4761 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.011081 4761 policy_none.go:49] "None policy: Start" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.012691 4761 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.012749 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.012805 4761 state_mem.go:35] "Initializing new in-memory state store" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.012841 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.027143 4761 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.081109 4761 manager.go:334] "Starting Device Plugin manager" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.081218 4761 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.081242 4761 server.go:79] "Starting device plugin registration server" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.081939 4761 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.081971 4761 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.082468 4761 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.082596 4761 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.082618 4761 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.092562 4761 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.110930 4761 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.111089 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.113118 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.113175 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.113188 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.114196 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.114318 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.115093 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116134 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116171 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116349 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116822 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.116855 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119654 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119685 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119750 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119775 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.120003 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.119935 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.120220 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.120812 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.120880 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.121685 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.123970 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124002 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124014 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124027 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124078 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124105 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124221 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124476 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.124589 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126254 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126306 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126411 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126469 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126493 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126909 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.126983 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.128200 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="400ms" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.128358 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.128416 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.128440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160130 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160192 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160250 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160287 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160322 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160479 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160589 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160640 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160676 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160806 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160847 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160881 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160914 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160948 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.160983 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.182631 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.184396 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.184459 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.184477 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.184516 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.185208 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262268 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262373 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262418 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262451 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262481 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262512 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262542 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262572 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262599 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262674 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262742 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262768 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262781 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262853 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262862 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262586 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262897 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262929 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262934 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262956 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.263028 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262970 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.263109 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262943 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262880 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262985 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262944 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.263339 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.262970 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.263494 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.386155 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.387725 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.387781 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.387800 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.387835 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.388257 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.451191 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.463895 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.486151 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.498352 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.507127 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.508436 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6c418909e11f029048c1fad4669ee97925941b1c360342bed699648ad44a3b36 WatchSource:0}: Error finding container 6c418909e11f029048c1fad4669ee97925941b1c360342bed699648ad44a3b36: Status 404 returned error can't find the container with id 6c418909e11f029048c1fad4669ee97925941b1c360342bed699648ad44a3b36 Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.509924 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-906a71ced1715106a54a10a93204dec9b285cf0dad8a604c4ffb6f63a9a43a80 WatchSource:0}: Error finding container 906a71ced1715106a54a10a93204dec9b285cf0dad8a604c4ffb6f63a9a43a80: Status 404 returned error can't find the container with id 906a71ced1715106a54a10a93204dec9b285cf0dad8a604c4ffb6f63a9a43a80 Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.515133 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-30ecb59df5a246a9efb61a0a021835bd954de15307c02752f7eb3437c8a5c8bb WatchSource:0}: Error finding container 30ecb59df5a246a9efb61a0a021835bd954de15307c02752f7eb3437c8a5c8bb: Status 404 returned error can't find the container with id 30ecb59df5a246a9efb61a0a021835bd954de15307c02752f7eb3437c8a5c8bb Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.521032 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-fd647a820e4884ae01088206713ad252305ab978a55238dfe316ff5e90b29ccf WatchSource:0}: Error finding container fd647a820e4884ae01088206713ad252305ab978a55238dfe316ff5e90b29ccf: Status 404 returned error can't find the container with id fd647a820e4884ae01088206713ad252305ab978a55238dfe316ff5e90b29ccf Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.529357 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="800ms" Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.533260 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-1822dbf780938e18c66825c6a5c7632206e8ab4b18ab4a9225b613b714531220 WatchSource:0}: Error finding container 1822dbf780938e18c66825c6a5c7632206e8ab4b18ab4a9225b613b714531220: Status 404 returned error can't find the container with id 1822dbf780938e18c66825c6a5c7632206e8ab4b18ab4a9225b613b714531220 Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.726951 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.727119 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.788507 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.790887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.790934 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.790949 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.790989 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.791392 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Nov 25 22:56:45 crc kubenswrapper[4761]: I1125 22:56:45.922620 4761 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:45 crc kubenswrapper[4761]: W1125 22:56:45.941264 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:45 crc kubenswrapper[4761]: E1125 22:56:45.941393 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.015107 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1822dbf780938e18c66825c6a5c7632206e8ab4b18ab4a9225b613b714531220"} Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.017072 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd647a820e4884ae01088206713ad252305ab978a55238dfe316ff5e90b29ccf"} Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.018680 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"30ecb59df5a246a9efb61a0a021835bd954de15307c02752f7eb3437c8a5c8bb"} Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.020062 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6c418909e11f029048c1fad4669ee97925941b1c360342bed699648ad44a3b36"} Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.022029 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"906a71ced1715106a54a10a93204dec9b285cf0dad8a604c4ffb6f63a9a43a80"} Nov 25 22:56:46 crc kubenswrapper[4761]: W1125 22:56:46.167386 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:46 crc kubenswrapper[4761]: E1125 22:56:46.167510 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:46 crc kubenswrapper[4761]: E1125 22:56:46.331285 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="1.6s" Nov 25 22:56:46 crc kubenswrapper[4761]: W1125 22:56:46.360958 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:46 crc kubenswrapper[4761]: E1125 22:56:46.361056 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.592323 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.594112 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.594168 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.594190 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.594234 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:46 crc kubenswrapper[4761]: E1125 22:56:46.594780 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Nov 25 22:56:46 crc kubenswrapper[4761]: I1125 22:56:46.923125 4761 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.028958 4761 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715" exitCode=0 Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.029093 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.029123 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.030461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.030526 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.030546 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.031070 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc" exitCode=0 Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.031155 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.031231 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.032417 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.032469 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.032486 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.034430 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.034448 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.034504 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.034387 4761 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785" exitCode=0 Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035554 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035804 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.035819 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.037448 4761 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08" exitCode=0 Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.037527 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.037621 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.038672 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.038744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.038762 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.050378 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.050478 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.050516 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.050545 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4"} Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.050620 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.074104 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.074153 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.074166 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:47 crc kubenswrapper[4761]: I1125 22:56:47.923032 4761 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:47 crc kubenswrapper[4761]: E1125 22:56:47.932938 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="3.2s" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.037190 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.046410 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.054583 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"13ea965844e67665a77c8ace18305a9ec605227d3b518ebc30d38454c627b216"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.054720 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.055764 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.055802 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.055812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.059322 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.059358 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.059371 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.059485 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.060810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.060864 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.060878 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.064134 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.064163 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.064176 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.064187 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.065621 4761 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5" exitCode=0 Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.065744 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.066152 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.066148 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5"} Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.066924 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.066969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.066981 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.068656 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.068737 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.068755 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:48 crc kubenswrapper[4761]: W1125 22:56:48.107325 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:48 crc kubenswrapper[4761]: E1125 22:56:48.107436 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:48 crc kubenswrapper[4761]: W1125 22:56:48.122177 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Nov 25 22:56:48 crc kubenswrapper[4761]: E1125 22:56:48.122267 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.195983 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.197480 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.197525 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.197545 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.197586 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:48 crc kubenswrapper[4761]: E1125 22:56:48.198174 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Nov 25 22:56:48 crc kubenswrapper[4761]: I1125 22:56:48.439446 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.073370 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3"} Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.074081 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.075332 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.075363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.075371 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076419 4761 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70" exitCode=0 Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076451 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70"} Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076519 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076584 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076596 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.076629 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077062 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077262 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077278 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077286 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077972 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078006 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.077983 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078046 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078066 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078069 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078097 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:49 crc kubenswrapper[4761]: I1125 22:56:49.078114 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.082981 4761 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083072 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.082878 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54"} Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083147 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665"} Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083174 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb"} Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083188 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd"} Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083005 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.083007 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084605 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084725 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084733 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084824 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.084840 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.470895 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:50 crc kubenswrapper[4761]: I1125 22:56:50.659781 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.092209 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.093063 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.093310 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c"} Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.093662 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094045 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094090 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094649 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.094666 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.095543 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.095585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.095601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.398638 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.400491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.400558 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.400585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:51 crc kubenswrapper[4761]: I1125 22:56:51.400629 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:56:52 crc kubenswrapper[4761]: I1125 22:56:52.094972 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:52 crc kubenswrapper[4761]: I1125 22:56:52.096167 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:52 crc kubenswrapper[4761]: I1125 22:56:52.096253 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:52 crc kubenswrapper[4761]: I1125 22:56:52.096280 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:52 crc kubenswrapper[4761]: I1125 22:56:52.496798 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.098381 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.099811 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.099864 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.099888 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.602975 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.603274 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.605420 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.605493 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.605520 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:53 crc kubenswrapper[4761]: I1125 22:56:53.796180 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:56:54 crc kubenswrapper[4761]: I1125 22:56:54.101802 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:54 crc kubenswrapper[4761]: I1125 22:56:54.103378 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:54 crc kubenswrapper[4761]: I1125 22:56:54.103421 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:54 crc kubenswrapper[4761]: I1125 22:56:54.103434 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:55 crc kubenswrapper[4761]: E1125 22:56:55.092750 4761 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 22:56:55 crc kubenswrapper[4761]: I1125 22:56:55.691793 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:55 crc kubenswrapper[4761]: I1125 22:56:55.691939 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:55 crc kubenswrapper[4761]: I1125 22:56:55.693026 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:55 crc kubenswrapper[4761]: I1125 22:56:55.693059 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:55 crc kubenswrapper[4761]: I1125 22:56:55.693070 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:57 crc kubenswrapper[4761]: I1125 22:56:57.263595 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 22:56:57 crc kubenswrapper[4761]: I1125 22:56:57.263909 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:57 crc kubenswrapper[4761]: I1125 22:56:57.265210 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:57 crc kubenswrapper[4761]: I1125 22:56:57.265270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:57 crc kubenswrapper[4761]: I1125 22:56:57.265284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:58 crc kubenswrapper[4761]: E1125 22:56:58.416606 4761 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187b61f9e27d99d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 22:56:44.920314328 +0000 UTC m=+0.653660193,LastTimestamp:2025-11-25 22:56:44.920314328 +0000 UTC m=+0.653660193,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.446590 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.446827 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.448195 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.448251 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.448270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:56:58 crc kubenswrapper[4761]: W1125 22:56:58.541556 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.541771 4761 trace.go:236] Trace[2102641877]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 22:56:48.539) (total time: 10001ms): Nov 25 22:56:58 crc kubenswrapper[4761]: Trace[2102641877]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (22:56:58.541) Nov 25 22:56:58 crc kubenswrapper[4761]: Trace[2102641877]: [10.001693305s] [10.001693305s] END Nov 25 22:56:58 crc kubenswrapper[4761]: E1125 22:56:58.541819 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.691875 4761 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.691987 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 22:56:58 crc kubenswrapper[4761]: I1125 22:56:58.924020 4761 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 22:56:59 crc kubenswrapper[4761]: W1125 22:56:59.511678 4761 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 22:56:59 crc kubenswrapper[4761]: I1125 22:56:59.511897 4761 trace.go:236] Trace[353669733]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 22:56:49.508) (total time: 10003ms): Nov 25 22:56:59 crc kubenswrapper[4761]: Trace[353669733]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10003ms (22:56:59.511) Nov 25 22:56:59 crc kubenswrapper[4761]: Trace[353669733]: [10.003780768s] [10.003780768s] END Nov 25 22:56:59 crc kubenswrapper[4761]: E1125 22:56:59.511943 4761 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 22:57:00 crc kubenswrapper[4761]: I1125 22:57:00.170452 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 22:57:00 crc kubenswrapper[4761]: I1125 22:57:00.170554 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 22:57:00 crc kubenswrapper[4761]: I1125 22:57:00.177207 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 22:57:00 crc kubenswrapper[4761]: I1125 22:57:00.177410 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.804001 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.804166 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.805636 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.805661 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.805672 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:03 crc kubenswrapper[4761]: I1125 22:57:03.812388 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.131275 4761 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.131329 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.132237 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.132272 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.132281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:04 crc kubenswrapper[4761]: I1125 22:57:04.625968 4761 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.092904 4761 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.164546 4761 trace.go:236] Trace[1727682620]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 22:56:54.253) (total time: 10910ms): Nov 25 22:57:05 crc kubenswrapper[4761]: Trace[1727682620]: ---"Objects listed" error: 10910ms (22:57:05.164) Nov 25 22:57:05 crc kubenswrapper[4761]: Trace[1727682620]: [10.91098057s] [10.91098057s] END Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.164590 4761 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.164650 4761 trace.go:236] Trace[760855883]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 22:56:54.024) (total time: 11139ms): Nov 25 22:57:05 crc kubenswrapper[4761]: Trace[760855883]: ---"Objects listed" error: 11139ms (22:57:05.164) Nov 25 22:57:05 crc kubenswrapper[4761]: Trace[760855883]: [11.139847369s] [11.139847369s] END Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.164661 4761 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.166401 4761 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.166985 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.167075 4761 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.202799 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49684->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.202874 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49698->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.202877 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49684->192.168.126.11:17697: read: connection reset by peer" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.202936 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49698->192.168.126.11:17697: read: connection reset by peer" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.203389 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.203420 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.267544 4761 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.703792 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.708781 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.922054 4761 apiserver.go:52] "Watching apiserver" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.928010 4761 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.928424 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.929018 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.929170 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.929305 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.929337 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.929400 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.929434 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.933160 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.933167 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.934002 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.934456 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:05 crc kubenswrapper[4761]: E1125 22:57:05.934536 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.935760 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.936015 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.937582 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.938188 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.938587 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.938857 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.939112 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.970940 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.986306 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:05 crc kubenswrapper[4761]: I1125 22:57:05.997499 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.014903 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.025050 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.026740 4761 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.039890 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.051854 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.063840 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071190 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071261 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071307 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071337 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071360 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071402 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071427 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071474 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071495 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071517 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071557 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071578 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071602 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071645 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071668 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071765 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071806 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071831 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071902 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071983 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072012 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072056 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072085 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072127 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072149 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072170 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072213 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072236 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072256 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072294 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072337 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072377 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072399 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072477 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072505 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072551 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072578 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072623 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072650 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072723 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072769 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071596 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071616 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071637 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071854 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.071870 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072081 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072453 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072457 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072510 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072579 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072608 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072721 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072791 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072891 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072793 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072963 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.072992 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073017 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073036 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073042 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073084 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073119 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073143 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073164 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073186 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073202 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073218 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073226 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073234 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073253 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073272 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073278 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073288 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073306 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073344 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073363 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073364 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073358 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073380 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073396 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073416 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073433 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073448 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073460 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073465 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073504 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073517 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073563 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073651 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073661 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073676 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073720 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073741 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073762 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073784 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073805 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073829 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073836 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073851 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073873 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073894 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073916 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073937 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073961 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073981 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.073983 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074013 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074032 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074050 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074101 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074101 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074118 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074136 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074139 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074151 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074194 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074509 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074666 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074954 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.074999 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075191 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075214 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075386 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075582 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075925 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076110 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076235 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076497 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076584 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076526 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076723 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.075936 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076777 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076884 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.076992 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.077078 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:06.577027786 +0000 UTC m=+22.310373621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077118 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077290 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077808 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077835 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077882 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.077905 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078101 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078128 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078150 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078173 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078195 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078217 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078238 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078259 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078280 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078302 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078324 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078345 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078365 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078385 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078405 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078425 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078447 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078470 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078494 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078514 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078535 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078555 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078576 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078597 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078626 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078646 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078665 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078686 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078727 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078748 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078768 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078791 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078816 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078838 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078858 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078879 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078901 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078920 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078941 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078968 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.078990 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079010 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079033 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079055 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079074 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079098 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079118 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079137 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079156 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079177 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079199 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079219 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079238 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079258 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079280 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079300 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079321 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079341 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079361 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079381 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079404 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079427 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079450 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079471 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079493 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079512 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079529 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079554 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079570 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079586 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079602 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079617 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079633 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079649 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079667 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079682 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079714 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079733 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079750 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079764 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079780 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079796 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079811 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079827 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079843 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079858 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079894 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079911 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079928 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079942 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079958 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079980 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.079996 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080011 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080026 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080044 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080060 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080077 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080094 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080110 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080152 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080181 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080209 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080233 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080253 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080274 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080291 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080308 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080327 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080344 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080363 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080386 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080409 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080428 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080483 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080493 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080503 4761 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080513 4761 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080522 4761 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080531 4761 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080540 4761 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080549 4761 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080558 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080567 4761 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080577 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080586 4761 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080595 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080605 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080615 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080624 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080633 4761 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080645 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080659 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080671 4761 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080682 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080794 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080806 4761 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080826 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.080857 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081130 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081300 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081434 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081452 4761 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081463 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081473 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081483 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081493 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081503 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081513 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081523 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081533 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081553 4761 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081562 4761 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081571 4761 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081581 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081591 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081601 4761 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081635 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081644 4761 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081654 4761 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081664 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081673 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081682 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081775 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.081787 4761 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.083169 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.083466 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.083670 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.083708 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.084071 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.084115 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.084133 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.084383 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.084746 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.085006 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.085038 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.085345 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.085651 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.086067 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.086122 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.086563 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.086875 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.086933 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087216 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087221 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087402 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087430 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087570 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.087944 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.088323 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.088796 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.090868 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091026 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091155 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091406 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091655 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091774 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.091916 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.092280 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.092551 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.093012 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.093846 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.094197 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.094390 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.094435 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.095201 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.095573 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.096099 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.096269 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.096396 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.096603 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.096820 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.096841 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.096920 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:06.596901406 +0000 UTC m=+22.330247241 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097002 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097433 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.097570 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097674 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.097681 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:06.597661785 +0000 UTC m=+22.331007720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097776 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097979 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098033 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097985 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.097578 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098173 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098228 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098359 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098390 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098713 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098774 4761 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.098958 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.099032 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100359 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100614 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100618 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100735 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100818 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100921 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.100951 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.101058 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.101187 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.101404 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.102708 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.102816 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.102837 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.102865 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.102997 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.103208 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.103399 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.103904 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104057 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104157 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104271 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104421 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104438 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104566 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.104853 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105000 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105293 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105440 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105346 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105549 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105581 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.105805 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.106008 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.106135 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.106506 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.107931 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.108166 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.121293 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.123941 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.129804 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.131817 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.132753 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.132866 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.134067 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.134902 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.134944 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.135849 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.141130 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.141799 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.156723 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157241 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157278 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157359 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157576 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157673 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157805 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157861 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157847 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.157914 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158306 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.158353 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158376 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158406 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158522 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158539 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.158559 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.159177 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.159483 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.159549 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:06.658480178 +0000 UTC m=+22.391826013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.159577 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:06.659563765 +0000 UTC m=+22.392909600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.160274 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.162468 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.163665 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3" exitCode=255 Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.163953 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3"} Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.167772 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.171544 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.171590 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.171790 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.171811 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.171868 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172021 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172030 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172141 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172260 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172313 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172343 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.172434 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.173021 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.173099 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.173266 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.173986 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.174157 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.174221 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.174530 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.174867 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.175091 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.176218 4761 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.177770 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.177775 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.177883 4761 scope.go:117] "RemoveContainer" containerID="0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.178686 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.179613 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182017 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182199 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182235 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182319 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182343 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182361 4761 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182358 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182371 4761 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182431 4761 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182442 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182451 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182461 4761 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182471 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182485 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182514 4761 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182522 4761 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182531 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182539 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182548 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182557 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182566 4761 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182592 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.182603 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183021 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183038 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183048 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183108 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183120 4761 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183129 4761 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183138 4761 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183147 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183174 4761 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183185 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183193 4761 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183202 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183211 4761 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183220 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183230 4761 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183257 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183265 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183273 4761 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183283 4761 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183292 4761 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183301 4761 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183328 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183338 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183346 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183356 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183368 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183379 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183414 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183428 4761 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183440 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183451 4761 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183462 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183499 4761 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183512 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183524 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183537 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183546 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183576 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183586 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183594 4761 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183603 4761 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183611 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183620 4761 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183629 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183656 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183665 4761 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183675 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183684 4761 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183724 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183735 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183743 4761 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183753 4761 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183761 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183779 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183813 4761 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183826 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183838 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183849 4761 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183861 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183898 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183909 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183921 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183933 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183944 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183982 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.183995 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184009 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184020 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184057 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184070 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184083 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184093 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184103 4761 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184132 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184145 4761 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184156 4761 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184165 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184177 4761 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184188 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184264 4761 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184300 4761 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184314 4761 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184328 4761 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184339 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184350 4761 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184382 4761 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184394 4761 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184407 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184419 4761 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184431 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184461 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184471 4761 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184482 4761 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184492 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184503 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184541 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184553 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184564 4761 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184576 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184632 4761 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184644 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184653 4761 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184662 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184671 4761 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184679 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184717 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184965 4761 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184977 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184985 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.184992 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185000 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185008 4761 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185036 4761 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185045 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185053 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185062 4761 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185071 4761 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185080 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185089 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185116 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185125 4761 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185133 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185141 4761 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185150 4761 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185159 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185166 4761 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185193 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.185203 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.190063 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.196541 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.200804 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.210035 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.219449 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.228800 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.235884 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.255200 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 22:57:06 crc kubenswrapper[4761]: W1125 22:57:06.268352 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6c430784e53538c2027fa11ac429a2669f6bec1c803aa9f2d32af645ac536f96 WatchSource:0}: Error finding container 6c430784e53538c2027fa11ac429a2669f6bec1c803aa9f2d32af645ac536f96: Status 404 returned error can't find the container with id 6c430784e53538c2027fa11ac429a2669f6bec1c803aa9f2d32af645ac536f96 Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.269260 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.282098 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.286339 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:57:06 crc kubenswrapper[4761]: W1125 22:57:06.287684 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-66022ced6a46671c69c4a2740933e19967fda4dbc9c26356e6c12e5f80d848b2 WatchSource:0}: Error finding container 66022ced6a46671c69c4a2740933e19967fda4dbc9c26356e6c12e5f80d848b2: Status 404 returned error can't find the container with id 66022ced6a46671c69c4a2740933e19967fda4dbc9c26356e6c12e5f80d848b2 Nov 25 22:57:06 crc kubenswrapper[4761]: W1125 22:57:06.310669 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6ea484dd49e069c55e358152e667e336de03ee5e191b9ec2a910d7012eb5955d WatchSource:0}: Error finding container 6ea484dd49e069c55e358152e667e336de03ee5e191b9ec2a910d7012eb5955d: Status 404 returned error can't find the container with id 6ea484dd49e069c55e358152e667e336de03ee5e191b9ec2a910d7012eb5955d Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.576007 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6qz4k"] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.576293 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.577777 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.577805 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.580331 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.587963 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.588116 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:07.588095153 +0000 UTC m=+23.321440988 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.591818 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.604794 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.612769 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.624035 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.634351 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.645592 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.654479 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.660485 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.670873 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689194 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/295217a2-b16b-4c48-ae61-00bfa20803fd-hosts-file\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689235 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmrh9\" (UniqueName: \"kubernetes.io/projected/295217a2-b16b-4c48-ae61-00bfa20803fd-kube-api-access-wmrh9\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689270 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689301 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689328 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.689355 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689401 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689424 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689486 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689497 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689530 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689540 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689506 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689603 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689490 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:07.689467637 +0000 UTC m=+23.422813472 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689685 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:07.689675702 +0000 UTC m=+23.423021627 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689721 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:07.689712293 +0000 UTC m=+23.423058248 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: E1125 22:57:06.689740 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:07.689733694 +0000 UTC m=+23.423079649 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.790117 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmrh9\" (UniqueName: \"kubernetes.io/projected/295217a2-b16b-4c48-ae61-00bfa20803fd-kube-api-access-wmrh9\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.790510 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/295217a2-b16b-4c48-ae61-00bfa20803fd-hosts-file\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.790884 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/295217a2-b16b-4c48-ae61-00bfa20803fd-hosts-file\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.809768 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmrh9\" (UniqueName: \"kubernetes.io/projected/295217a2-b16b-4c48-ae61-00bfa20803fd-kube-api-access-wmrh9\") pod \"node-resolver-6qz4k\" (UID: \"295217a2-b16b-4c48-ae61-00bfa20803fd\") " pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.894677 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6qz4k" Nov 25 22:57:06 crc kubenswrapper[4761]: W1125 22:57:06.943168 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod295217a2_b16b_4c48_ae61_00bfa20803fd.slice/crio-3b469f16c2e2d7f8b487f317186c0e396c3f3df28fdfc1c1064e8a49b32673cf WatchSource:0}: Error finding container 3b469f16c2e2d7f8b487f317186c0e396c3f3df28fdfc1c1064e8a49b32673cf: Status 404 returned error can't find the container with id 3b469f16c2e2d7f8b487f317186c0e396c3f3df28fdfc1c1064e8a49b32673cf Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.974552 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-7g447"] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.975265 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-fnqp9"] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.975443 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4lp9m"] Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.975759 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.976109 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.976429 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fnqp9" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.977850 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.978785 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.979155 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.979317 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.979445 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.979577 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.980126 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.980271 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.980455 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.980989 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.981102 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.981251 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 22:57:06 crc kubenswrapper[4761]: I1125 22:57:06.996333 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:06Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.010494 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.010659 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.015942 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.016432 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.017726 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.018358 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.019364 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.019876 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.020504 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.021408 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.022024 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.022904 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.023458 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.024481 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.024981 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.025256 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.025476 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.026373 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.026902 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.027798 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.028166 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.028722 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.029720 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.030172 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.031182 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.031602 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.032592 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.033045 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.033608 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.034665 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.035219 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.036101 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.036568 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.037423 4761 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.037544 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.039298 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.040214 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.040637 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.042070 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.043187 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.043364 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.044060 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.044727 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.045793 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.046345 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.047471 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.048149 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.049064 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.049496 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.050352 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.050846 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.051889 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.052344 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.053165 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.053606 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.054474 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.055023 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.055464 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.062078 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.076910 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.093620 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.094937 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-kubelet\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.094991 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-binary-copy\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095046 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-proxy-tls\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095095 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-os-release\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095130 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfl7z\" (UniqueName: \"kubernetes.io/projected/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-kube-api-access-lfl7z\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095162 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-os-release\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095294 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-rootfs\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095363 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-mcd-auth-proxy-config\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095392 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgsnx\" (UniqueName: \"kubernetes.io/projected/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-kube-api-access-kgsnx\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095421 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-multus\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095442 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-hostroot\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095476 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-socket-dir-parent\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095498 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-conf-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095517 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-multus-certs\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095535 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095584 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095666 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-cnibin\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095780 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-etc-kubernetes\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095850 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cni-binary-copy\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095893 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-netns\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095925 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-system-cni-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.095966 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-system-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096014 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-bin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096080 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096114 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-k8s-cni-cncf-io\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096225 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cnibin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096261 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-daemon-config\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.096289 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vvj9\" (UniqueName: \"kubernetes.io/projected/fda38150-e7bb-458c-8a8e-476ce0804735-kube-api-access-9vvj9\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.121345 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.137838 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.152345 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.168218 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.170252 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.170841 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.173415 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6qz4k" event={"ID":"295217a2-b16b-4c48-ae61-00bfa20803fd","Type":"ContainerStarted","Data":"3b469f16c2e2d7f8b487f317186c0e396c3f3df28fdfc1c1064e8a49b32673cf"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.176866 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.176917 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.176931 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6ea484dd49e069c55e358152e667e336de03ee5e191b9ec2a910d7012eb5955d"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.178212 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.179292 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6c430784e53538c2027fa11ac429a2669f6bec1c803aa9f2d32af645ac536f96"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.181052 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.181126 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"66022ced6a46671c69c4a2740933e19967fda4dbc9c26356e6c12e5f80d848b2"} Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.191511 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196647 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-socket-dir-parent\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196674 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-conf-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196705 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-multus-certs\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196723 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196739 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196754 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-cnibin\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196770 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-etc-kubernetes\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196786 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cni-binary-copy\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196785 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-conf-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196822 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-multus-certs\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196841 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-netns\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196843 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-cnibin\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196862 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-etc-kubernetes\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196878 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-socket-dir-parent\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196801 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-netns\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196920 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196955 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-system-cni-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.196936 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-system-cni-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197032 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-system-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197064 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-bin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197088 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197115 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-system-cni-dir\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197116 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-k8s-cni-cncf-io\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197148 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-run-k8s-cni-cncf-io\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197177 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-bin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197453 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197148 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-daemon-config\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197538 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vvj9\" (UniqueName: \"kubernetes.io/projected/fda38150-e7bb-458c-8a8e-476ce0804735-kube-api-access-9vvj9\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197574 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cnibin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197596 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-kubelet\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197611 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-binary-copy\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197635 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-proxy-tls\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197637 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cni-binary-copy\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197674 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-os-release\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197764 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-os-release\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197787 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfl7z\" (UniqueName: \"kubernetes.io/projected/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-kube-api-access-lfl7z\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197814 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-rootfs\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197837 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-mcd-auth-proxy-config\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197861 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgsnx\" (UniqueName: \"kubernetes.io/projected/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-kube-api-access-kgsnx\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197874 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-multus-daemon-config\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197816 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197839 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-kubelet\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197922 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-hostroot\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197946 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fda38150-e7bb-458c-8a8e-476ce0804735-os-release\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197978 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-os-release\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198051 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-cnibin\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198151 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-rootfs\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.197888 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-hostroot\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198193 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-multus\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198245 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-host-var-lib-cni-multus\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198531 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-mcd-auth-proxy-config\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.198536 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fda38150-e7bb-458c-8a8e-476ce0804735-cni-binary-copy\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.203831 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-proxy-tls\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.206737 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.214416 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfl7z\" (UniqueName: \"kubernetes.io/projected/5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6-kube-api-access-lfl7z\") pod \"multus-fnqp9\" (UID: \"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\") " pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.216922 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgsnx\" (UniqueName: \"kubernetes.io/projected/7e27d0af-be2c-4fc2-93f7-2f5294537b9d-kube-api-access-kgsnx\") pod \"machine-config-daemon-4lp9m\" (UID: \"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\") " pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.220924 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vvj9\" (UniqueName: \"kubernetes.io/projected/fda38150-e7bb-458c-8a8e-476ce0804735-kube-api-access-9vvj9\") pod \"multus-additional-cni-plugins-7g447\" (UID: \"fda38150-e7bb-458c-8a8e-476ce0804735\") " pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.221370 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.233121 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.278548 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.295268 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.300650 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 22:57:07 crc kubenswrapper[4761]: W1125 22:57:07.311629 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e27d0af_be2c_4fc2_93f7_2f5294537b9d.slice/crio-58cdfb3f89966ee378e342d1258ec466b1b07bbc40fb12c6c85974f8419f2f2d WatchSource:0}: Error finding container 58cdfb3f89966ee378e342d1258ec466b1b07bbc40fb12c6c85974f8419f2f2d: Status 404 returned error can't find the container with id 58cdfb3f89966ee378e342d1258ec466b1b07bbc40fb12c6c85974f8419f2f2d Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.312829 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7g447" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.313326 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.320215 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.321106 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.323569 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fnqp9" Nov 25 22:57:07 crc kubenswrapper[4761]: W1125 22:57:07.326813 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfda38150_e7bb_458c_8a8e_476ce0804735.slice/crio-9e4e6a08a52dbf8068e27a1cfba5a0f2a8a3d9d8de495b5ab64bffe159a54f80 WatchSource:0}: Error finding container 9e4e6a08a52dbf8068e27a1cfba5a0f2a8a3d9d8de495b5ab64bffe159a54f80: Status 404 returned error can't find the container with id 9e4e6a08a52dbf8068e27a1cfba5a0f2a8a3d9d8de495b5ab64bffe159a54f80 Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.344748 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.358295 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qdk9h"] Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.358972 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.360821 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.361009 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.362298 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.362406 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.363264 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.363375 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.363319 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.363782 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.379912 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.392545 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.411297 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.427318 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.444312 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.459602 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.482334 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.495004 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503007 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503040 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503058 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lj29\" (UniqueName: \"kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503155 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503189 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503208 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503225 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503256 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503274 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503303 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503327 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503381 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503401 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503418 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503432 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503446 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503472 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503510 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503549 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.503577 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.508110 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.518876 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.529383 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.545641 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.557944 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.569552 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.580289 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.592793 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604124 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604425 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.604451 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:09.604418869 +0000 UTC m=+25.337764704 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604498 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604536 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604555 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604573 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604589 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604605 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604621 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604651 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604657 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604671 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604724 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604742 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604755 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lj29\" (UniqueName: \"kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604777 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604785 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604808 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604848 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604922 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604942 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.604958 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605024 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605047 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605060 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605072 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605094 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605111 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605135 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605150 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605189 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605222 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605251 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605276 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605337 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605367 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605519 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605572 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.605630 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.606440 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.620544 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.632738 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.644714 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.706238 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.706291 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.706311 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.706340 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706387 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706452 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706459 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706469 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706483 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706487 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:09.70646664 +0000 UTC m=+25.439812575 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706490 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706522 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:09.706503751 +0000 UTC m=+25.439849686 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706516 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706541 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:09.706533092 +0000 UTC m=+25.439879057 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706549 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:07 crc kubenswrapper[4761]: E1125 22:57:07.706602 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:09.706571613 +0000 UTC m=+25.439917548 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.735467 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.735641 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lj29\" (UniqueName: \"kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29\") pod \"ovnkube-node-qdk9h\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: I1125 22:57:07.969488 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:07 crc kubenswrapper[4761]: W1125 22:57:07.988470 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0ef866e_3c5b_4dfc_b5a2_61ee81e8bde8.slice/crio-df94ff5676095a41933e5f37b2147d17ac04659d8ce4fff4beb6357d980be2fc WatchSource:0}: Error finding container df94ff5676095a41933e5f37b2147d17ac04659d8ce4fff4beb6357d980be2fc: Status 404 returned error can't find the container with id df94ff5676095a41933e5f37b2147d17ac04659d8ce4fff4beb6357d980be2fc Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.010746 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.010778 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:08 crc kubenswrapper[4761]: E1125 22:57:08.010841 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:08 crc kubenswrapper[4761]: E1125 22:57:08.010913 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.184662 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" exitCode=0 Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.184746 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.184774 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"df94ff5676095a41933e5f37b2147d17ac04659d8ce4fff4beb6357d980be2fc"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.187175 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerStarted","Data":"e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.187205 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerStarted","Data":"363580ed2e1722aa6ff602fb669abdfa4b077a16169205791fcf1e6b1980bee7"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.190184 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52" exitCode=0 Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.190270 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.190324 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerStarted","Data":"9e4e6a08a52dbf8068e27a1cfba5a0f2a8a3d9d8de495b5ab64bffe159a54f80"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.191896 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.191925 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.191940 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"58cdfb3f89966ee378e342d1258ec466b1b07bbc40fb12c6c85974f8419f2f2d"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.202558 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6qz4k" event={"ID":"295217a2-b16b-4c48-ae61-00bfa20803fd","Type":"ContainerStarted","Data":"03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1"} Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.203163 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: E1125 22:57:08.211600 4761 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.216390 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.232489 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.264045 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.279807 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.292635 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.307624 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.321365 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.338540 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.352550 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.371504 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.383610 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.406190 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.418353 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.450909 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.463564 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.487538 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.500252 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.514118 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.524110 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.541958 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.555886 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.570441 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.584488 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.597274 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.609778 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.622629 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.635847 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.821178 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jvzpt"] Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.821867 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.825684 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.825915 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.825908 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.826164 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.839753 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.852202 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.870938 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.883002 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.896433 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.911953 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.917414 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbj5z\" (UniqueName: \"kubernetes.io/projected/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-kube-api-access-fbj5z\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.917442 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-serviceca\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.917482 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-host\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.933236 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.952362 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:08 crc kubenswrapper[4761]: I1125 22:57:08.967827 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:08Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.012112 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.012215 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.015794 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.033065 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-host\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.033144 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbj5z\" (UniqueName: \"kubernetes.io/projected/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-kube-api-access-fbj5z\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.033189 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-serviceca\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.034155 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-serviceca\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.034217 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-host\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.054489 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.064300 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbj5z\" (UniqueName: \"kubernetes.io/projected/d9744f71-40a6-41d3-a8a3-6b1a16ea5740-kube-api-access-fbj5z\") pod \"node-ca-jvzpt\" (UID: \"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\") " pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.068495 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.080056 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.110490 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.159256 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.208322 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jvzpt" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.211412 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218089 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218133 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218220 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218246 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218259 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.218273 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.221525 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6" exitCode=0 Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.222012 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6"} Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.223899 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.237665 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.272223 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.319130 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.352252 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.392143 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.432797 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.471973 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.513420 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.553276 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.592284 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.631078 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.640419 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.640596 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:13.640578172 +0000 UTC m=+29.373924017 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.680513 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.734840 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.741682 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.741796 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.741844 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.741881 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.741925 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.741984 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742004 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742023 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742047 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742065 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742070 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:13.742048248 +0000 UTC m=+29.475394113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742118 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:13.742100819 +0000 UTC m=+29.475446694 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742161 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742197 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742204 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:13.742191132 +0000 UTC m=+29.475536997 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: E1125 22:57:09.742251 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:13.742237813 +0000 UTC m=+29.475583678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.755985 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.800946 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.834493 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.880631 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.921509 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.956628 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:09 crc kubenswrapper[4761]: I1125 22:57:09.999665 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:09Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.010487 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:10 crc kubenswrapper[4761]: E1125 22:57:10.010680 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.010503 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:10 crc kubenswrapper[4761]: E1125 22:57:10.010849 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.052601 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.075610 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.119402 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.162177 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.196555 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.228236 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jvzpt" event={"ID":"d9744f71-40a6-41d3-a8a3-6b1a16ea5740","Type":"ContainerStarted","Data":"a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829"} Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.228303 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jvzpt" event={"ID":"d9744f71-40a6-41d3-a8a3-6b1a16ea5740","Type":"ContainerStarted","Data":"266f5fbae4daa8bab48741310dba8bb8db9e74b703a5b2d84a8d383965fe644e"} Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.233190 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9" exitCode=0 Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.233559 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9"} Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.255236 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.283640 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.319539 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.355474 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.402842 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.435295 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.472244 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.515012 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.564385 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.597992 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.638724 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.683148 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.723484 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.759532 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.798769 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.841657 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.876030 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.918189 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:10 crc kubenswrapper[4761]: I1125 22:57:10.959180 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:10Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.010652 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.010852 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.247652 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6" exitCode=0 Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.247755 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6"} Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.266905 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.288035 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.310750 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.329680 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.355941 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.380524 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.412836 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.427287 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.445314 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.465544 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.477898 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.498791 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.515029 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.531643 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.553884 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.568227 4761 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.570349 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.570398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.570420 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.570508 4761 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.585374 4761 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.585627 4761 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.586795 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.586867 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.586886 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.586911 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.586932 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.604682 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.609372 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.609406 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.609415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.609429 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.609439 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.624879 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.629896 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.629961 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.629978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.630002 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.630021 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.649538 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.654018 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.654076 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.654094 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.654119 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.654137 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.672049 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.676337 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.676394 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.676412 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.676438 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.676455 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.693977 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:11Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:11 crc kubenswrapper[4761]: E1125 22:57:11.694246 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.696027 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.696082 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.696100 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.696125 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.696144 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.799270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.799328 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.799344 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.799368 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.799416 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.906641 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.906756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.906782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.906811 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:11 crc kubenswrapper[4761]: I1125 22:57:11.906833 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:11Z","lastTransitionTime":"2025-11-25T22:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010084 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010147 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010165 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010207 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010274 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:12 crc kubenswrapper[4761]: E1125 22:57:12.010475 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.010598 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:12 crc kubenswrapper[4761]: E1125 22:57:12.010760 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.112956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.113017 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.113029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.113048 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.113061 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.215907 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.216195 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.216329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.216463 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.216615 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.254948 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434" exitCode=0 Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.255028 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.261601 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.287133 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.303417 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.319797 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.319873 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.319901 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.319932 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.319953 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.326554 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.345106 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.359975 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.380715 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.407236 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.423809 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.423855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.423866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.423883 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.423895 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.425131 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.444389 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.460693 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.485941 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.505503 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.525276 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.526862 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.526893 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.526904 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.526921 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.526932 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.542166 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.555872 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:12Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.630049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.630112 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.630124 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.630142 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.630153 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.732442 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.732507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.732544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.732581 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.732605 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.835968 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.836032 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.836049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.836076 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.836094 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.938941 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.939004 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.939020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.939040 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:12 crc kubenswrapper[4761]: I1125 22:57:12.939053 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:12Z","lastTransitionTime":"2025-11-25T22:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.010900 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.011043 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.041435 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.041509 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.041534 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.041566 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.041590 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.144893 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.145247 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.145259 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.145276 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.145289 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.248673 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.248759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.248778 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.248801 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.248817 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.269059 4761 generic.go:334] "Generic (PLEG): container finished" podID="fda38150-e7bb-458c-8a8e-476ce0804735" containerID="c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395" exitCode=0 Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.269117 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerDied","Data":"c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.294957 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.313625 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.331204 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.352499 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.353666 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.353737 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.353749 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.353784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.353795 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.366163 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.386811 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.399259 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.414733 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.434039 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.456936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.456971 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.456980 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.457016 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.457028 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.461202 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.480723 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.540935 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.556883 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.559670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.559759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.559776 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.559803 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.559821 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.575313 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.599980 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.662365 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.662402 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.662413 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.662430 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.662441 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.686977 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.687171 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.687145126 +0000 UTC m=+37.420490971 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.764897 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.764959 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.764978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.765001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.765019 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.788570 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.788617 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.788673 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788745 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.788759 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788820 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.788797757 +0000 UTC m=+37.522143622 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788869 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788953 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.78892682 +0000 UTC m=+37.522272695 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788971 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789032 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789054 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789142 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.789123466 +0000 UTC m=+37.522469311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.788969 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789238 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789266 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:13 crc kubenswrapper[4761]: E1125 22:57:13.789354 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.789331991 +0000 UTC m=+37.522677896 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.868328 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.868385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.868398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.868418 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.868430 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.971216 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.971273 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.971292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.971320 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:13 crc kubenswrapper[4761]: I1125 22:57:13.971338 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:13Z","lastTransitionTime":"2025-11-25T22:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.010187 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.010216 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:14 crc kubenswrapper[4761]: E1125 22:57:14.010367 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:14 crc kubenswrapper[4761]: E1125 22:57:14.010468 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.073916 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.073980 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.073992 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.074008 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.074019 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.177130 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.177189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.177205 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.177230 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.177247 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.277017 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" event={"ID":"fda38150-e7bb-458c-8a8e-476ce0804735","Type":"ContainerStarted","Data":"d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.278987 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.279034 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.279050 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.279069 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.279083 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.283541 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.284123 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.284174 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.295299 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.311461 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.325617 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.326633 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.341731 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.358258 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.373751 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.381951 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.382023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.382034 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.382070 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.382083 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.386229 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.404213 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.419262 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.447905 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.483319 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.484994 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.485047 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.485064 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.485086 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.485103 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.512668 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.533324 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.549174 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.559859 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.578657 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.587583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.587676 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.587771 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.587835 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.587894 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.589965 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.601193 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.613117 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.625272 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.643255 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.658891 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.671269 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.689653 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.689735 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.689754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.689779 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.689795 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.691136 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.709264 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.720465 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.731784 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.749931 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.764962 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.787621 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:14Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.792550 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.792599 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.792611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.792630 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.792641 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.894605 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.895029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.895181 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.895325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.895471 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.998042 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.998108 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.998127 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.998156 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:14 crc kubenswrapper[4761]: I1125 22:57:14.998175 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:14Z","lastTransitionTime":"2025-11-25T22:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.010901 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:15 crc kubenswrapper[4761]: E1125 22:57:15.011124 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.039106 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.059023 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.079655 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.103432 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.103488 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.103507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.103541 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.103559 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.104416 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.119511 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.148271 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.169247 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.189643 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206393 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206458 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206456 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206505 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.206618 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.225527 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.243401 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.259168 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.280881 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.287078 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.299178 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.309659 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.309735 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.309753 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.309777 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.309798 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.313636 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.320684 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.335739 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.355039 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.373546 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.390107 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.412493 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.412554 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.412580 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.412613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.412633 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.415316 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.433509 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.466883 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.487845 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.512716 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.514794 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.514833 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.514845 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.514861 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.514871 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.533074 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.547994 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.573465 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.596507 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.616940 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.617645 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.617738 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.617757 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.617785 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.617806 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.639007 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.720962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.721012 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.721023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.721041 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.721051 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.824024 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.824067 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.824074 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.824089 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.824097 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.926921 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.926962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.926973 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.926989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:15 crc kubenswrapper[4761]: I1125 22:57:15.926998 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:15Z","lastTransitionTime":"2025-11-25T22:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.010540 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.010611 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:16 crc kubenswrapper[4761]: E1125 22:57:16.010662 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:16 crc kubenswrapper[4761]: E1125 22:57:16.010736 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.031647 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.031674 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.031682 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.031709 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.031717 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.134618 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.134675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.134692 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.134741 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.134759 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.236794 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.236839 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.236848 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.236863 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.236875 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.338867 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.338901 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.338909 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.338922 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.338931 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.441910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.442005 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.442035 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.442102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.442130 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.544553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.544584 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.544592 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.544605 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.544615 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.648164 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.648237 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.648259 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.648291 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.648313 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.751510 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.751549 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.751561 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.751591 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.751602 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.854620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.854660 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.854671 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.854685 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.854716 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.957579 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.957647 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.957665 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.957694 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:16 crc kubenswrapper[4761]: I1125 22:57:16.957746 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:16Z","lastTransitionTime":"2025-11-25T22:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.011165 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:17 crc kubenswrapper[4761]: E1125 22:57:17.011326 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.061954 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.061989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.061998 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.062011 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.062019 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.165440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.165485 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.165496 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.165513 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.165524 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.269294 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.269348 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.269365 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.269391 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.269410 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.302277 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/0.log" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.307355 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8" exitCode=1 Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.307417 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.308375 4761 scope.go:117] "RemoveContainer" containerID="4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.372545 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.372601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.372619 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.372646 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.372667 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.476059 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.476116 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.476133 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.476158 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.476175 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.541783 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.563369 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.579490 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.579931 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.579943 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.579767 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.579962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.580140 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.602220 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:16Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 22:57:16.662450 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:57:16.662493 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 22:57:16.662540 6073 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 22:57:16.662593 6073 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:16.662620 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:16.662627 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:16.662646 6073 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:16.662765 6073 factory.go:656] Stopping watch factory\\\\nI1125 22:57:16.662791 6073 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:16.662824 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:16.662839 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:16.662864 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:16.662877 6073 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:16.662889 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:16.662900 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.620564 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.639553 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.659137 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.675143 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.682820 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.682861 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.682873 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.682891 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.682903 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.690473 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.706027 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.724397 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.742230 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.756559 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.769969 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.780718 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:17Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.785359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.785405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.785421 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.785441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.785456 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.887800 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.887842 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.887855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.887870 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.887881 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.992217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.992261 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.992272 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.992290 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:17 crc kubenswrapper[4761]: I1125 22:57:17.992304 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:17Z","lastTransitionTime":"2025-11-25T22:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.010816 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.010827 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:18 crc kubenswrapper[4761]: E1125 22:57:18.010946 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:18 crc kubenswrapper[4761]: E1125 22:57:18.011034 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.094497 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.094557 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.094574 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.094596 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.094613 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.197622 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.197686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.197730 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.197756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.197774 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.300800 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.300863 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.300882 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.300906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.300925 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.314877 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/0.log" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.319448 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.320002 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.343033 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.362278 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.377647 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.403441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.403485 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.403496 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.403514 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.403526 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.404514 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:16Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 22:57:16.662450 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:57:16.662493 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 22:57:16.662540 6073 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 22:57:16.662593 6073 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:16.662620 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:16.662627 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:16.662646 6073 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:16.662765 6073 factory.go:656] Stopping watch factory\\\\nI1125 22:57:16.662791 6073 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:16.662824 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:16.662839 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:16.662864 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:16.662877 6073 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:16.662889 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:16.662900 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.425497 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.440011 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.454863 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.467575 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.480343 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.499505 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.506734 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.506938 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.507021 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.507102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.507177 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.518933 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.532833 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.551314 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.567802 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.602242 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:18Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.610302 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.610363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.610381 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.610406 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.610425 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.713259 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.713321 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.713337 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.713360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.713378 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.816048 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.816105 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.816117 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.816136 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.816149 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.919834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.919896 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.919913 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.919936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:18 crc kubenswrapper[4761]: I1125 22:57:18.919958 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:18Z","lastTransitionTime":"2025-11-25T22:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.011036 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:19 crc kubenswrapper[4761]: E1125 22:57:19.011329 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.022044 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.022088 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.022104 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.022130 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.022144 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.125274 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.125326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.125338 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.125357 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.125370 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.228487 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.228542 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.228560 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.228583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.228598 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.326127 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/1.log" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.327303 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/0.log" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.330720 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.330744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.330752 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.330765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.330775 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.331318 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94" exitCode=1 Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.331359 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.331399 4761 scope.go:117] "RemoveContainer" containerID="4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.332763 4761 scope.go:117] "RemoveContainer" containerID="c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94" Nov 25 22:57:19 crc kubenswrapper[4761]: E1125 22:57:19.333018 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.350617 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.364303 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.375382 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.401385 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:16Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 22:57:16.662450 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:57:16.662493 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 22:57:16.662540 6073 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 22:57:16.662593 6073 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:16.662620 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:16.662627 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:16.662646 6073 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:16.662765 6073 factory.go:656] Stopping watch factory\\\\nI1125 22:57:16.662791 6073 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:16.662824 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:16.662839 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:16.662864 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:16.662877 6073 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:16.662889 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:16.662900 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.418251 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.433352 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.433403 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.433415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.433430 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.433440 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.434297 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.449785 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.462241 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.478397 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.490559 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.503600 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.516551 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.535978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.535979 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.536066 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.536092 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.536122 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.536139 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.569724 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.589922 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:19Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.639367 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.639423 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.639441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.639467 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.639485 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.742142 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.742182 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.742193 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.742209 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.742246 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.845007 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.845412 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.845623 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.845878 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.846128 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.949549 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.949909 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.950081 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.950281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:19 crc kubenswrapper[4761]: I1125 22:57:19.950420 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:19Z","lastTransitionTime":"2025-11-25T22:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.010340 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.010462 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:20 crc kubenswrapper[4761]: E1125 22:57:20.010574 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:20 crc kubenswrapper[4761]: E1125 22:57:20.011308 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.053768 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.053832 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.053848 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.053873 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.053890 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.156380 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.156443 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.156465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.156491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.156509 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.209941 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7"] Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.210604 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.213305 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.213842 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.232425 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.254447 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.258846 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.258894 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.258910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.258933 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.258950 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.272078 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.286228 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.303656 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.321195 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.337057 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/1.log" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.339653 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.341258 4761 scope.go:117] "RemoveContainer" containerID="c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94" Nov 25 22:57:20 crc kubenswrapper[4761]: E1125 22:57:20.341472 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.356464 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f76fd05-689a-44c0-9a1b-019930a01198-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.356534 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.356571 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n5cp\" (UniqueName: \"kubernetes.io/projected/0f76fd05-689a-44c0-9a1b-019930a01198-kube-api-access-4n5cp\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.356594 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.361028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.361086 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.361104 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.361127 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.361144 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.364437 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.382341 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.413571 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.431293 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.443353 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.455662 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.457199 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f76fd05-689a-44c0-9a1b-019930a01198-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.457346 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.457380 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n5cp\" (UniqueName: \"kubernetes.io/projected/0f76fd05-689a-44c0-9a1b-019930a01198-kube-api-access-4n5cp\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.457405 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.458135 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.459074 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f76fd05-689a-44c0-9a1b-019930a01198-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.463801 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f76fd05-689a-44c0-9a1b-019930a01198-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.464507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.464543 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.464555 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.464572 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.464584 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.472264 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.479799 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n5cp\" (UniqueName: \"kubernetes.io/projected/0f76fd05-689a-44c0-9a1b-019930a01198-kube-api-access-4n5cp\") pod \"ovnkube-control-plane-749d76644c-xppl7\" (UID: \"0f76fd05-689a-44c0-9a1b-019930a01198\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.481160 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.497224 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea9f9e9dcb5bf1c4b49306665b44c2fdb4876d23e05ecc2713a183018cfcba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:16Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 22:57:16.662450 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:57:16.662493 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 22:57:16.662540 6073 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 22:57:16.662593 6073 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:16.662620 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:16.662627 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:16.662646 6073 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:16.662765 6073 factory.go:656] Stopping watch factory\\\\nI1125 22:57:16.662791 6073 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:16.662824 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:16.662839 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:16.662864 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:16.662877 6073 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:16.662889 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:16.662900 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.509103 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.520968 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.539156 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.543910 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.562162 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.566645 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.566684 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.566722 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.566817 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.566840 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.578045 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.589928 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.618456 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.634245 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.652954 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669332 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669388 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669417 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.669867 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.691957 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.708229 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.721775 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.740652 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.759427 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.769402 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.772618 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.772688 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.772735 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.772759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.772776 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.785883 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:20Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.875190 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.876995 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.877020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.877038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.877051 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.980049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.980128 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.980151 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.980182 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.980205 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:20Z","lastTransitionTime":"2025-11-25T22:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.981880 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lhpwb"] Nov 25 22:57:20 crc kubenswrapper[4761]: I1125 22:57:20.982532 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:20 crc kubenswrapper[4761]: E1125 22:57:20.982638 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.002801 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.010724 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.010842 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.024371 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.036559 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.053579 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.071842 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.082803 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.082854 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.082866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.082886 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.082898 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.085587 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.107888 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.119153 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.129160 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.148982 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.165177 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.165532 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl798\" (UniqueName: \"kubernetes.io/projected/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-kube-api-access-bl798\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.165649 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.178174 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.185038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.185075 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.185087 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.185103 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.185115 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.192907 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.206613 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.223161 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.264271 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.266816 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl798\" (UniqueName: \"kubernetes.io/projected/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-kube-api-access-bl798\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.266901 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.267027 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.267101 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:21.767082722 +0000 UTC m=+37.500428557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.288028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.288073 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.288085 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.288101 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.288111 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.291884 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl798\" (UniqueName: \"kubernetes.io/projected/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-kube-api-access-bl798\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.301219 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.345477 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" event={"ID":"0f76fd05-689a-44c0-9a1b-019930a01198","Type":"ContainerStarted","Data":"7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.345531 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" event={"ID":"0f76fd05-689a-44c0-9a1b-019930a01198","Type":"ContainerStarted","Data":"fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.345544 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" event={"ID":"0f76fd05-689a-44c0-9a1b-019930a01198","Type":"ContainerStarted","Data":"ab7dadeeae08b56ba146fbcdd451866cb2e9c03532030fc650f484856948a0ef"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.364534 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.379913 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.390860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.390906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.390917 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.390936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.390947 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.396761 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.412201 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.423577 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.434051 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.447659 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.460880 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.472980 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.493254 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.493308 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.493325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.493348 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.493366 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.494486 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.509376 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.525946 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.541929 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.554517 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.570083 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.595794 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.595858 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.595875 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.595899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.595917 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.603270 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.620448 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.698606 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.698681 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.698744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.698772 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.698791 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.772006 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.772189 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:57:37.772151607 +0000 UTC m=+53.505497452 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.772350 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.772569 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.772670 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:22.77264507 +0000 UTC m=+38.505990945 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.801482 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.801539 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.801558 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.801583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.801606 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.873154 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.873236 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.873308 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.873380 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873440 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873488 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873530 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873581 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873613 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873647 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873560 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:37.873532931 +0000 UTC m=+53.606878796 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873685 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873734 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873770 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:37.873744697 +0000 UTC m=+53.607090562 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873796 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:37.873784658 +0000 UTC m=+53.607130523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.873820 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:37.873808108 +0000 UTC m=+53.607153973 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.902381 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.902447 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.902467 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.902495 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.902515 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.923552 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.935414 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.935503 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.935530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.935563 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.935591 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.959751 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.964503 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.964574 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.964592 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.964616 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.964633 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:21 crc kubenswrapper[4761]: E1125 22:57:21.981956 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:21Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.986988 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.987038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.987056 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.987080 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:21 crc kubenswrapper[4761]: I1125 22:57:21.987098 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:21Z","lastTransitionTime":"2025-11-25T22:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.007341 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:22Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.010588 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.010666 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.010756 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.010951 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.011074 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.011225 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.012969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.013374 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.013737 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.013895 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.014022 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.037961 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:22Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.038272 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.041054 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.041143 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.041162 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.041187 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.041204 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.145050 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.145317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.145440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.145560 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.145672 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.250119 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.250185 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.250203 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.250226 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.250245 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.353107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.353172 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.353189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.353217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.353236 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.456801 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.456875 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.456900 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.456931 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.456953 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.560167 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.560230 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.560247 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.560271 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.560289 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.663507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.663901 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.664081 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.664468 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.664859 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.768803 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.768860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.768876 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.768900 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.768924 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.782684 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.782915 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:22 crc kubenswrapper[4761]: E1125 22:57:22.783023 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:24.782996463 +0000 UTC m=+40.516342328 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.872845 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.872895 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.872903 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.872920 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.872932 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.975996 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.976052 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.976069 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.976094 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:22 crc kubenswrapper[4761]: I1125 22:57:22.976116 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:22Z","lastTransitionTime":"2025-11-25T22:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.011215 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:23 crc kubenswrapper[4761]: E1125 22:57:23.011742 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.079254 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.079322 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.079340 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.079365 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.079382 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.182292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.182355 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.182373 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.182425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.182445 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.284943 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.284989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.285002 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.285020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.285031 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.388306 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.388364 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.388380 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.388406 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.388425 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.491633 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.491687 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.491719 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.491740 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.491756 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.594612 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.594656 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.594674 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.594739 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.594761 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.697030 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.697327 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.697416 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.697514 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.697596 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.800918 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.801226 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.801387 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.801511 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.801636 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.904639 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.904721 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.904738 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.904761 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:23 crc kubenswrapper[4761]: I1125 22:57:23.904778 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:23Z","lastTransitionTime":"2025-11-25T22:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.008075 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.008146 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.008169 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.008195 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.008217 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.010690 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.010764 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.010692 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:24 crc kubenswrapper[4761]: E1125 22:57:24.010926 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:24 crc kubenswrapper[4761]: E1125 22:57:24.011061 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:24 crc kubenswrapper[4761]: E1125 22:57:24.011147 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.111439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.111597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.111624 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.111652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.111673 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.214575 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.214641 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.214658 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.214683 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.214736 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.318398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.318733 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.318866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.318989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.319157 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.422438 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.422520 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.422540 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.422582 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.422617 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.525845 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.525882 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.525892 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.525906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.525917 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.628334 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.628370 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.628379 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.628393 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.628402 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.731455 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.731539 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.731562 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.731593 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.731615 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.803259 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:24 crc kubenswrapper[4761]: E1125 22:57:24.803513 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:24 crc kubenswrapper[4761]: E1125 22:57:24.803620 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:28.803598516 +0000 UTC m=+44.536944441 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.834398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.834461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.834479 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.834535 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.834551 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.938036 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.938085 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.938098 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.938115 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:24 crc kubenswrapper[4761]: I1125 22:57:24.938125 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:24Z","lastTransitionTime":"2025-11-25T22:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.010188 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:25 crc kubenswrapper[4761]: E1125 22:57:25.010373 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.029277 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.040713 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.040749 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.040760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.040777 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.040788 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.048735 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.061474 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.081620 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.098398 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.118199 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.133168 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.143458 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.143511 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.143529 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.143553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.143569 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.151688 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.175657 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.193794 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.213796 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.230542 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.245503 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.247769 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.247810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.247827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.247849 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.247866 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.262604 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.280380 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.297239 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.323146 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:25Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.350840 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.350902 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.350920 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.350944 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.350961 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.454140 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.454192 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.454209 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.454236 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.454254 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.580537 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.580575 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.580583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.580595 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.580604 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.683688 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.683769 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.683785 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.683808 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.683823 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.786786 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.786850 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.786870 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.786897 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.786915 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.890775 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.890888 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.890899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.890924 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.890937 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.993829 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.994270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.994299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.994330 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:25 crc kubenswrapper[4761]: I1125 22:57:25.994356 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:25Z","lastTransitionTime":"2025-11-25T22:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.010678 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.010769 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.010781 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:26 crc kubenswrapper[4761]: E1125 22:57:26.010865 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:26 crc kubenswrapper[4761]: E1125 22:57:26.010960 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:26 crc kubenswrapper[4761]: E1125 22:57:26.011182 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.097398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.097468 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.097498 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.097523 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.097540 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.204112 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.204162 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.204174 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.204197 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.204209 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.307320 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.307446 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.307469 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.307496 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.307513 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.410750 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.410832 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.410851 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.410878 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.410899 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.514384 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.514439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.514456 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.514482 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.514500 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.618467 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.619384 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.619528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.619731 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.619893 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.723769 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.723826 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.723844 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.723868 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.723885 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.827106 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.827205 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.827233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.827261 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.827283 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.930775 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.930818 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.930834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.930856 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:26 crc kubenswrapper[4761]: I1125 22:57:26.930873 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:26Z","lastTransitionTime":"2025-11-25T22:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.010383 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:27 crc kubenswrapper[4761]: E1125 22:57:27.010841 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.033007 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.033065 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.033077 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.033094 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.033106 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.136584 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.136646 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.136665 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.136690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.136737 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.240575 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.240654 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.240672 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.240750 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.240776 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.343907 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.344082 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.344113 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.344147 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.344171 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.449890 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.449951 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.449971 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.449995 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.450013 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.553418 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.553494 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.553507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.553533 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.553549 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.657244 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.657318 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.657336 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.657363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.657381 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.761167 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.761239 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.761257 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.761286 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.761315 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.864378 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.864409 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.864417 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.864430 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.864440 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.967105 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.967171 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.967191 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.967215 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:27 crc kubenswrapper[4761]: I1125 22:57:27.967235 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:27Z","lastTransitionTime":"2025-11-25T22:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.010500 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.010553 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:28 crc kubenswrapper[4761]: E1125 22:57:28.010675 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.010694 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:28 crc kubenswrapper[4761]: E1125 22:57:28.010852 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:28 crc kubenswrapper[4761]: E1125 22:57:28.010956 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.070677 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.070747 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.070759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.070779 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.070791 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.174637 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.174815 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.174842 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.174877 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.174897 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.277217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.277463 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.277570 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.277645 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.277731 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.380870 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.380952 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.380975 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.381006 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.381031 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.484230 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.484285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.484294 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.484314 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.484326 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.588175 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.588236 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.588247 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.588275 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.588289 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.691161 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.691223 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.691241 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.691264 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.691284 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.794538 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.794619 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.794638 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.794663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.794681 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.865375 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:28 crc kubenswrapper[4761]: E1125 22:57:28.865730 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:28 crc kubenswrapper[4761]: E1125 22:57:28.865909 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:36.865868752 +0000 UTC m=+52.599214617 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.898009 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.898143 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.898163 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.898186 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:28 crc kubenswrapper[4761]: I1125 22:57:28.898203 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:28Z","lastTransitionTime":"2025-11-25T22:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.000781 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.000825 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.000843 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.000865 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.000881 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.010496 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:29 crc kubenswrapper[4761]: E1125 22:57:29.010830 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.104755 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.104822 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.104839 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.104866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.104885 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.207637 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.207735 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.207754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.207780 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.207799 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.341982 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.342047 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.342065 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.342089 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.342108 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.445214 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.445283 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.445302 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.445330 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.445348 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.552608 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.552684 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.552756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.552784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.552803 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.656519 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.656595 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.656620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.656652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.656675 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.759742 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.759813 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.759832 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.759856 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.759875 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.862285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.862335 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.862351 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.862372 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.862386 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.965690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.965782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.965799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.965823 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:29 crc kubenswrapper[4761]: I1125 22:57:29.965839 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:29Z","lastTransitionTime":"2025-11-25T22:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.010386 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.010458 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.010387 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:30 crc kubenswrapper[4761]: E1125 22:57:30.010604 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:30 crc kubenswrapper[4761]: E1125 22:57:30.010870 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:30 crc kubenswrapper[4761]: E1125 22:57:30.010982 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.069421 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.069463 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.069473 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.069487 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.069496 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.172759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.172848 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.172872 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.172904 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.172928 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.276072 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.276155 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.276173 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.276200 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.276219 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.379553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.379626 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.379650 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.379678 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.379733 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.483640 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.483762 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.483790 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.483844 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.483873 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.586947 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.587002 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.587013 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.587026 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.587035 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.690138 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.690198 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.690212 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.690231 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.690247 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.793217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.793284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.793300 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.793324 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.793342 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.896599 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.896654 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.896671 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.896692 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.896740 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.999272 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.999405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.999424 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.999449 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:30 crc kubenswrapper[4761]: I1125 22:57:30.999467 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:30Z","lastTransitionTime":"2025-11-25T22:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.010996 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:31 crc kubenswrapper[4761]: E1125 22:57:31.011173 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.102120 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.102176 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.102194 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.102211 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.102223 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.205248 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.205314 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.205324 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.205338 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.205346 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.309199 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.309273 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.309293 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.309326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.309363 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.413039 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.413111 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.413135 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.413166 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.413187 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.515904 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.515978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.516000 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.516029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.516073 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.619652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.619756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.619773 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.619797 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.619817 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.722730 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.722793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.722811 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.722866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.722884 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.825824 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.825877 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.825894 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.825919 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.825937 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.929797 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.929879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.929902 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.929932 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:31 crc kubenswrapper[4761]: I1125 22:57:31.929954 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:31Z","lastTransitionTime":"2025-11-25T22:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.010140 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.010247 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.010326 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.010351 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.010489 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.010626 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.032538 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.032621 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.032644 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.032674 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.032731 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.065628 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.066081 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.066116 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.066143 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.066161 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.085414 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:32Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.090473 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.090544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.090561 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.090586 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.090605 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.111927 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:32Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.117330 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.117398 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.117441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.117475 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.117499 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.138443 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:32Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.144297 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.144359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.144377 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.144401 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.144418 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.169621 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:32Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.174900 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.174947 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.174963 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.174983 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.175001 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.195795 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:32Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:32 crc kubenswrapper[4761]: E1125 22:57:32.196017 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.198597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.198672 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.198693 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.198738 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.198754 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.301601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.301665 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.301686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.301756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.301777 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.404168 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.404581 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.404823 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.405043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.405191 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.508869 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.508927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.508946 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.508971 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.508989 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.611831 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.611968 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.611998 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.612028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.612051 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.715258 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.715317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.715334 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.715357 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.715377 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.818425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.818655 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.818691 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.818776 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.818803 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.922367 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.922435 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.922457 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.922487 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:32 crc kubenswrapper[4761]: I1125 22:57:32.922510 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:32Z","lastTransitionTime":"2025-11-25T22:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.011018 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:33 crc kubenswrapper[4761]: E1125 22:57:33.011226 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.025810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.025866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.025885 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.025909 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.025926 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.129371 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.129431 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.129448 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.129471 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.129490 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.232589 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.232658 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.232680 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.232732 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.232750 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.335579 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.335664 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.335688 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.335765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.335789 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.438742 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.438797 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.438814 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.438837 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.438854 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.542132 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.542201 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.542219 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.542243 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.542261 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.645280 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.645356 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.645375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.645402 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.645423 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.748747 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.748812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.748831 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.748858 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.748876 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.852760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.852837 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.852861 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.852892 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.852914 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.957323 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.957375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.957391 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.957415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:33 crc kubenswrapper[4761]: I1125 22:57:33.957433 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:33Z","lastTransitionTime":"2025-11-25T22:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.011050 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.011092 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:34 crc kubenswrapper[4761]: E1125 22:57:34.011217 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.011136 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:34 crc kubenswrapper[4761]: E1125 22:57:34.011404 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:34 crc kubenswrapper[4761]: E1125 22:57:34.011557 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.060913 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.061326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.061506 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.061680 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.061907 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.166300 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.167029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.167094 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.167132 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.167157 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.270437 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.270502 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.270519 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.270544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.270562 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.373886 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.373968 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.373989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.374023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.374049 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.477402 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.477478 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.477504 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.477533 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.477555 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.581236 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.581368 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.581388 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.581419 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.581440 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.684501 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.684603 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.684625 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.684659 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.684681 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.788257 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.788317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.788336 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.788363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.788380 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.891417 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.891537 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.891561 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.891590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.891613 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.995126 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.995184 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.995200 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.995223 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:34 crc kubenswrapper[4761]: I1125 22:57:34.995239 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:34Z","lastTransitionTime":"2025-11-25T22:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.010610 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:35 crc kubenswrapper[4761]: E1125 22:57:35.010825 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.012810 4761 scope.go:117] "RemoveContainer" containerID="c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.033441 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.058469 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.076740 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.098572 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.098910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.099120 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.099255 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.099387 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.105614 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.127473 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.143835 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.175654 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.194690 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.203054 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.203278 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.203440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.203592 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.203780 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.210854 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.230796 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.250391 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.266150 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.296688 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.306670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.306741 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.306759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.306781 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.306800 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.314161 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.336368 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.355467 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.378396 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.402875 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/1.log" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.405901 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.406907 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.408785 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.408906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.408977 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.409053 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.409075 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.423317 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.443860 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.465923 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.481642 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.493134 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.512254 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.513360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.513465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.513488 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.513513 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.513565 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.531638 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.547593 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.577071 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.600822 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616377 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616549 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616574 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.616627 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.661037 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.681069 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.702519 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.718540 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.719528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.719564 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.719576 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.719593 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.719607 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.734578 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.747671 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:35Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.822333 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.822404 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.822424 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.822447 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.822464 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.925675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.925769 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.925790 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.925817 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:35 crc kubenswrapper[4761]: I1125 22:57:35.925836 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:35Z","lastTransitionTime":"2025-11-25T22:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.010214 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.010294 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.010213 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.010458 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.010607 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.010822 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.030189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.030252 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.030303 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.030335 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.030354 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.133480 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.133528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.133540 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.133558 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.133570 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.236452 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.236518 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.236530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.236555 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.236567 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.339368 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.339452 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.339467 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.339495 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.339513 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.412558 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/2.log" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.413652 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/1.log" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.418358 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" exitCode=1 Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.418439 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.418565 4761 scope.go:117] "RemoveContainer" containerID="c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.419618 4761 scope.go:117] "RemoveContainer" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.419980 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.442370 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.442443 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.442461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.442487 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.442503 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.454299 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.472423 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.490381 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.510221 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.533237 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.545006 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.545064 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.545082 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.545106 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.545124 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.552535 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.586454 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.604976 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.625030 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.646364 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.648472 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.648553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.648585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.648616 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.648636 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.662280 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.678401 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.696284 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.713395 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.735416 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.751995 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.752024 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.752031 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.752049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.752059 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.754320 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.768782 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:36Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.854692 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.854753 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.854763 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.854777 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.854788 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.873087 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.873317 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:36 crc kubenswrapper[4761]: E1125 22:57:36.873482 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:57:52.873445693 +0000 UTC m=+68.606791558 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.957818 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.957880 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.957899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.957927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:36 crc kubenswrapper[4761]: I1125 22:57:36.957948 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:36Z","lastTransitionTime":"2025-11-25T22:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.011176 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.011398 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.061476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.061543 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.061563 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.061587 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.061606 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.164786 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.164865 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.164883 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.165346 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.165403 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.179485 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.194269 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.195805 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.214641 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.232070 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.243382 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.268821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.268927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.268943 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.268960 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.268972 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.273476 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7872fb43448394f61995ef0a937276ebcc196ab933b2891baac3ea57ff35b94\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:18Z\\\",\\\"message\\\":\\\" for removal\\\\nI1125 22:57:18.406388 6189 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:18.406394 6189 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:18.406419 6189 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 22:57:18.406443 6189 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 22:57:18.406458 6189 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 22:57:18.406466 6189 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:18.406475 6189 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:57:18.406494 6189 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:18.406509 6189 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 22:57:18.406518 6189 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 22:57:18.406536 6189 factory.go:656] Stopping watch factory\\\\nI1125 22:57:18.406551 6189 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:18.406565 6189 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 22:57:18.406572 6189 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 22:57:18.406578 6189 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 22:57:18.406588 6189 handler.go:208] Removed *v1.Pod event handler 6\\\\nF1125 22:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.294376 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.313548 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.332372 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.348245 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.363927 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.371837 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.371875 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.371887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.371906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.371918 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.381821 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.395775 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.407767 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.423524 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/2.log" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.427000 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.429078 4761 scope.go:117] "RemoveContainer" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.429237 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.454910 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.470285 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.474124 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.474187 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.474199 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.474217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.474229 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.485950 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.515684 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.531540 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.548199 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.561105 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.576622 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.576690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.576736 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.576762 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.576785 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.591253 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.610053 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.628753 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.649760 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.669680 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.679528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.679561 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.679571 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.679585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.679594 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.691268 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.709221 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.725943 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.743986 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.759832 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.778782 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.782965 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.783343 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:58:09.783267193 +0000 UTC m=+85.516613068 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.783419 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.783464 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.783480 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.783504 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.783523 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.789468 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.806809 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.827883 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:37Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.884192 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.884232 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.884340 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.884376 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884376 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884448 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884454 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884511 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884527 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884524 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884555 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884573 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884457 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:58:09.884438462 +0000 UTC m=+85.617784317 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884607 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:58:09.884589036 +0000 UTC m=+85.617934881 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884625 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:58:09.884616256 +0000 UTC m=+85.617962101 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:37 crc kubenswrapper[4761]: E1125 22:57:37.884639 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:58:09.884631757 +0000 UTC m=+85.617977602 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.885812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.885855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.885868 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.885887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.885900 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.992795 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.992838 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.992864 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.992879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:37 crc kubenswrapper[4761]: I1125 22:57:37.992889 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:37Z","lastTransitionTime":"2025-11-25T22:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.010505 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:38 crc kubenswrapper[4761]: E1125 22:57:38.010618 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.010675 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:38 crc kubenswrapper[4761]: E1125 22:57:38.010741 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.010782 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:38 crc kubenswrapper[4761]: E1125 22:57:38.010820 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.095617 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.095666 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.095678 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.095717 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.095730 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.198599 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.198654 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.198670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.198722 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.198741 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.301756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.301807 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.301821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.301839 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.301852 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.404249 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.404304 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.404322 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.404344 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.404363 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.506565 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.506624 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.506642 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.506665 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.506681 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.608941 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.609001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.609023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.609049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.609070 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.716955 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.717002 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.717013 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.717033 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.717043 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.819578 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.819632 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.819644 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.819662 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.819673 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.923277 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.923359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.923383 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.923414 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:38 crc kubenswrapper[4761]: I1125 22:57:38.923439 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:38Z","lastTransitionTime":"2025-11-25T22:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.010429 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:39 crc kubenswrapper[4761]: E1125 22:57:39.010619 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.026115 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.026180 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.026196 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.026217 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.026232 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.129849 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.129928 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.129948 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.129973 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.129991 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.233271 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.233773 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.233801 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.233827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.233849 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.336891 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.336930 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.336941 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.336957 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.336968 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.438849 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.438898 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.438914 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.438936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.438952 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.542175 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.542246 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.542271 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.542303 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.542325 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.645362 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.645415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.645432 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.645453 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.645469 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.748000 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.748062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.748081 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.748106 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.748125 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.851460 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.851515 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.851532 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.851556 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.851574 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.954657 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.954760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.954786 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.954814 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:39 crc kubenswrapper[4761]: I1125 22:57:39.954833 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:39Z","lastTransitionTime":"2025-11-25T22:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.010929 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.011005 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.011011 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:40 crc kubenswrapper[4761]: E1125 22:57:40.011145 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:40 crc kubenswrapper[4761]: E1125 22:57:40.011319 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:40 crc kubenswrapper[4761]: E1125 22:57:40.011453 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.057856 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.057909 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.057926 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.057953 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.057970 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.161738 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.161794 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.161810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.161835 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.161853 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.264902 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.264969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.264987 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.265014 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.265032 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.368427 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.368488 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.368505 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.368529 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.368548 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.471416 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.471465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.471482 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.471507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.471524 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.574333 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.574401 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.574418 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.574441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.574460 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.677506 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.677566 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.677583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.677607 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.677626 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.780860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.780930 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.780947 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.780974 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.780993 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.884023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.884081 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.884098 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.884124 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.884143 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.987173 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.987259 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.987279 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.987343 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:40 crc kubenswrapper[4761]: I1125 22:57:40.987370 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:40Z","lastTransitionTime":"2025-11-25T22:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.011010 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:41 crc kubenswrapper[4761]: E1125 22:57:41.011208 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.089681 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.089771 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.089788 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.089813 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.089830 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.193231 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.193298 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.193314 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.193342 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.193364 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.296318 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.296424 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.296451 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.296485 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.296542 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.399425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.399488 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.399505 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.399530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.399548 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.502334 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.502389 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.502405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.502427 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.502443 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.605528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.605598 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.605621 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.605656 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.605679 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.708840 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.708907 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.708925 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.708952 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.708970 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.812249 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.812326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.812356 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.812389 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.812417 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.914917 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.914985 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.915001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.915026 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:41 crc kubenswrapper[4761]: I1125 22:57:41.915044 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:41Z","lastTransitionTime":"2025-11-25T22:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.010686 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.010789 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.010686 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.010896 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.011232 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.011330 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.018342 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.018418 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.018443 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.018474 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.018500 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.122037 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.122098 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.122116 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.122139 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.122158 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.224727 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.225023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.225040 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.225063 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.225083 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.297673 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.297754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.297772 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.297793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.297812 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.317846 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:42Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.322962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.323035 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.323059 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.323085 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.323101 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.343092 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:42Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.348283 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.348344 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.348363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.348386 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.348403 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.368184 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:42Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.373104 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.373164 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.373181 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.373204 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.373221 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.393206 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:42Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.398029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.398094 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.398117 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.398148 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.398168 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.419338 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:42Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:42 crc kubenswrapper[4761]: E1125 22:57:42.419596 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.421914 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.421964 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.421983 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.422004 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.422021 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.525691 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.525804 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.525826 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.525860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.525882 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.629324 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.629408 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.629431 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.629461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.629484 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.732674 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.732771 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.732810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.732836 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.732854 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.836355 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.836435 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.836461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.836491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.836516 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.939813 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.939906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.939940 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.939971 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:42 crc kubenswrapper[4761]: I1125 22:57:42.939991 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:42Z","lastTransitionTime":"2025-11-25T22:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.010959 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:43 crc kubenswrapper[4761]: E1125 22:57:43.011169 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.042600 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.042664 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.042681 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.042746 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.042767 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.146125 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.146212 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.146251 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.146283 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.146308 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.249945 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.250011 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.250028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.250051 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.250068 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.352611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.352668 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.352685 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.352737 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.352760 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.461219 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.461299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.461320 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.461345 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.461365 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.564346 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.564409 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.564430 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.564455 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.564472 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.667669 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.667778 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.667798 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.667825 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.667843 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.770787 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.770853 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.770870 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.770894 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.770913 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.873292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.873351 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.873367 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.873389 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.873405 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.976630 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.976685 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.976729 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.976754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:43 crc kubenswrapper[4761]: I1125 22:57:43.976771 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:43Z","lastTransitionTime":"2025-11-25T22:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.010218 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.010277 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:44 crc kubenswrapper[4761]: E1125 22:57:44.010500 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.010530 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:44 crc kubenswrapper[4761]: E1125 22:57:44.010654 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:44 crc kubenswrapper[4761]: E1125 22:57:44.010848 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.079831 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.079901 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.079955 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.079986 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.080009 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.183102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.183153 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.183169 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.183193 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.183209 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.286226 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.286284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.286300 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.286324 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.286341 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.389338 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.389390 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.389408 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.389429 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.389445 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.493194 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.493270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.493293 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.493322 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.493340 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.597079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.597140 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.597160 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.597185 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.597207 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.700597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.700662 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.700686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.700753 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.700778 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.803021 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.803070 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.803084 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.803107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.803119 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.906169 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.906233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.906245 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.906263 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:44 crc kubenswrapper[4761]: I1125 22:57:44.906276 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:44Z","lastTransitionTime":"2025-11-25T22:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.009271 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.009325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.009358 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.009385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.009399 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.011313 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:45 crc kubenswrapper[4761]: E1125 22:57:45.011656 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.034638 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.051940 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.071760 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.090153 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.112481 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.112531 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.112548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.112571 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.112588 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.115584 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.131916 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.163866 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.182082 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.197984 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.214051 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.215022 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.215061 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.215076 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.215096 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.215111 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.231586 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.250218 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.270101 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.284527 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.316452 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.322188 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.322240 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.322257 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.322281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.322300 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.339208 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.358913 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.373313 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:45Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.425906 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.425969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.425986 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.426012 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.426032 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.528879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.529250 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.529405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.529655 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.529880 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.633455 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.633512 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.633529 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.633553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.633570 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.740363 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.740434 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.740468 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.740515 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.740541 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.843801 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.843848 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.843865 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.843888 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.843904 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.946863 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.947316 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.947595 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.947911 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:45 crc kubenswrapper[4761]: I1125 22:57:45.948420 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:45Z","lastTransitionTime":"2025-11-25T22:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.010663 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.011467 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.012072 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:46 crc kubenswrapper[4761]: E1125 22:57:46.012176 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:46 crc kubenswrapper[4761]: E1125 22:57:46.012330 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:46 crc kubenswrapper[4761]: E1125 22:57:46.012486 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.052123 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.052186 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.052209 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.052238 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.052263 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.154954 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.155030 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.155066 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.155098 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.155120 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.257774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.257827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.257845 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.257868 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.257884 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.360609 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.360675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.360728 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.360759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.360782 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.463531 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.463626 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.463653 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.463738 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.463767 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.567038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.567147 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.567172 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.567202 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.567224 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.670308 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.670608 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.670633 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.670663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.670685 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.773113 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.773173 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.773197 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.773218 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.773231 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.875421 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.875455 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.875465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.875476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.875485 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.977841 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.977899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.977915 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.977939 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:46 crc kubenswrapper[4761]: I1125 22:57:46.977958 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:46Z","lastTransitionTime":"2025-11-25T22:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.010552 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:47 crc kubenswrapper[4761]: E1125 22:57:47.010753 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.081125 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.081190 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.081209 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.081233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.081250 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.183793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.183864 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.183881 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.183905 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.183921 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.287001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.287060 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.287079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.287100 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.287117 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.389495 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.389563 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.389582 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.389612 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.389631 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.493235 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.493291 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.493310 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.493333 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.493352 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.596320 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.596373 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.596388 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.596412 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.596429 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.699950 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.700017 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.700037 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.700064 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.700089 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.802753 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.802816 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.802835 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.802861 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.802881 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.905046 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.905125 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.905149 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.905180 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:47 crc kubenswrapper[4761]: I1125 22:57:47.905205 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:47Z","lastTransitionTime":"2025-11-25T22:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.008461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.008552 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.008580 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.008622 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.008652 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.010689 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.010817 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.010689 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:48 crc kubenswrapper[4761]: E1125 22:57:48.010903 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:48 crc kubenswrapper[4761]: E1125 22:57:48.011080 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:48 crc kubenswrapper[4761]: E1125 22:57:48.011231 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.111953 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.112043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.112200 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.112283 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.112324 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.215171 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.215260 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.215285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.215315 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.215337 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.318914 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.319047 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.319074 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.319104 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.319122 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.422552 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.422883 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.422902 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.422926 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.422943 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.526692 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.526826 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.526850 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.526874 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.526892 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.629634 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.629686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.629733 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.629757 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.629774 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.732766 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.732903 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.732925 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.732956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.732981 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.835688 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.835800 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.835818 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.835840 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.835857 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.938161 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.938230 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.938253 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.938285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:48 crc kubenswrapper[4761]: I1125 22:57:48.938307 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:48Z","lastTransitionTime":"2025-11-25T22:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.010818 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:49 crc kubenswrapper[4761]: E1125 22:57:49.010995 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.041192 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.041260 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.041277 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.041299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.041316 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.144448 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.144521 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.144545 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.144602 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.144626 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.247132 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.247209 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.247235 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.247265 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.247286 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.350072 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.350141 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.350159 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.350183 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.350203 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.454107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.454181 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.454200 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.454229 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.454248 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.556815 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.556872 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.556886 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.556905 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.556923 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.659392 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.659449 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.659466 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.659490 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.659506 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.762836 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.762888 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.762899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.762915 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.762927 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.865574 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.865635 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.865650 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.865671 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.865687 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.969070 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.969135 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.969159 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.969191 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:49 crc kubenswrapper[4761]: I1125 22:57:49.969214 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:49Z","lastTransitionTime":"2025-11-25T22:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.010962 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.010977 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.011113 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:50 crc kubenswrapper[4761]: E1125 22:57:50.011254 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:50 crc kubenswrapper[4761]: E1125 22:57:50.011394 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:50 crc kubenswrapper[4761]: E1125 22:57:50.011468 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.072181 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.072273 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.072293 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.072318 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.072336 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.175360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.175420 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.175440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.175466 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.175496 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.278110 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.278158 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.278169 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.278187 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.278200 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.380279 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.380319 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.380329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.380360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.380370 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.481671 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.481746 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.481764 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.481785 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.481804 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.583544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.583575 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.583584 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.583596 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.583855 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.685880 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.685940 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.685956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.685978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.685995 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.787944 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.787988 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.788003 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.788020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.788033 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.890222 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.890280 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.890299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.890323 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.890341 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.993565 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.993640 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.993663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.993690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:50 crc kubenswrapper[4761]: I1125 22:57:50.993741 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:50Z","lastTransitionTime":"2025-11-25T22:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.011878 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:51 crc kubenswrapper[4761]: E1125 22:57:51.012078 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.012722 4761 scope.go:117] "RemoveContainer" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" Nov 25 22:57:51 crc kubenswrapper[4761]: E1125 22:57:51.012959 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.096673 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.096761 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.096784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.096809 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.096828 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.198165 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.198205 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.198222 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.198238 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.198250 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.301274 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.301326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.301338 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.301356 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.301369 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.404538 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.404624 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.404640 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.404664 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.404681 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.508196 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.508267 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.508284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.508309 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.508329 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.611611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.611686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.611697 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.611724 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.611733 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.714766 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.714880 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.714898 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.714921 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.714939 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.818077 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.818168 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.818193 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.818226 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.818250 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.921072 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.921118 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.921126 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.921140 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:51 crc kubenswrapper[4761]: I1125 22:57:51.921149 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:51Z","lastTransitionTime":"2025-11-25T22:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.038179 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.038197 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.038426 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.038307 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.038301 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.038533 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.039754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.039786 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.039799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.039816 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.039828 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.143052 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.143096 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.143112 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.143135 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.143153 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.245855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.245928 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.245942 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.245963 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.245976 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.348665 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.348787 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.348813 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.348844 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.348862 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.452201 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.452281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.452314 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.452343 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.452364 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.535652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.535717 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.535728 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.535742 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.535753 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.549430 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:52Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.552978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.553038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.553056 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.553080 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.553100 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.574031 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:52Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.579548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.579608 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.579622 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.579638 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.579648 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.600190 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:52Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.606656 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.606751 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.606775 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.606799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.606845 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.627791 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:52Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.633469 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.633538 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.633556 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.633581 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.633594 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.650277 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:52Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.650449 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.652279 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.652317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.652329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.652347 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.652361 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.755652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.755775 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.755802 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.755831 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.755853 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.858619 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.858666 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.858681 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.858709 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.858756 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.893570 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.893821 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:52 crc kubenswrapper[4761]: E1125 22:57:52.893956 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:58:24.893929195 +0000 UTC m=+100.627275060 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.961607 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.961662 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.961678 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.961733 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:52 crc kubenswrapper[4761]: I1125 22:57:52.961751 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:52Z","lastTransitionTime":"2025-11-25T22:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.010971 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:53 crc kubenswrapper[4761]: E1125 22:57:53.011128 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.064009 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.064049 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.064059 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.064072 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.064082 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.165960 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.166001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.166012 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.166030 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.166041 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.268675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.268779 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.268810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.268834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.268854 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.371693 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.371774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.371791 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.371813 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.371832 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.474226 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.474270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.474286 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.474306 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.474323 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.577788 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.577843 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.577862 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.577883 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.577900 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.680476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.680518 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.680529 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.680546 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.680559 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.783933 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.783992 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.784009 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.784074 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.784092 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.886594 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.886648 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.886659 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.886676 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.886687 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.988992 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.989029 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.989041 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.989057 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:53 crc kubenswrapper[4761]: I1125 22:57:53.989067 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:53Z","lastTransitionTime":"2025-11-25T22:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.010797 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.010848 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:54 crc kubenswrapper[4761]: E1125 22:57:54.010920 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.010971 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:54 crc kubenswrapper[4761]: E1125 22:57:54.011162 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:54 crc kubenswrapper[4761]: E1125 22:57:54.011679 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.091670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.091706 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.091735 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.091752 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.091760 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.194544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.194830 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.194840 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.194854 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.194865 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.297601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.297644 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.297653 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.297667 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.297676 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.400113 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.400175 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.400244 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.400269 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.400299 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.503408 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.503456 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.503466 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.503481 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.503490 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.508367 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/0.log" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.508410 4761 generic.go:334] "Generic (PLEG): container finished" podID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" containerID="e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f" exitCode=1 Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.508447 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerDied","Data":"e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.508918 4761 scope.go:117] "RemoveContainer" containerID="e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.527668 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.541438 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.561158 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.581932 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.602071 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.607219 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.607268 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.607283 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.607304 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.607319 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.614158 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.630075 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.645800 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.656773 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.668039 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.680832 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.695011 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.706155 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.712716 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.712742 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.712751 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.712764 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.712775 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.724538 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.737503 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.753569 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.771997 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.786746 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:54Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.814712 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.814747 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.814757 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.814773 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.814783 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.916669 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.916709 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.916732 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.916744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:54 crc kubenswrapper[4761]: I1125 22:57:54.916752 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:54Z","lastTransitionTime":"2025-11-25T22:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.010690 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:55 crc kubenswrapper[4761]: E1125 22:57:55.011939 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.019026 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.019062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.019073 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.019085 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.019094 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.027973 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.043645 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.061928 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.076048 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.106306 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122795 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122831 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122841 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122865 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.122859 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.137919 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.153475 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.166296 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.177999 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.189331 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.200518 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.216559 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.225528 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.225579 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.225594 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.225615 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.225634 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.231095 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.241190 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.265678 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.279295 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.301216 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.328887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.328952 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.328963 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.328982 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.328995 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.431757 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.431825 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.431836 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.431859 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.431875 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.514318 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/0.log" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.514405 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerStarted","Data":"aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.534321 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.534385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.534403 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.534433 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.534452 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.543817 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.558722 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.574066 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.590927 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.605097 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.619441 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.632036 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.637461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.637499 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.637509 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.637525 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.637536 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.660895 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.675142 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.693167 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.709991 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.724949 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.739980 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.740027 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.740039 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.740057 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.740070 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.740530 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.757068 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.770164 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.786279 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.801261 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.813679 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:57:55Z is after 2025-08-24T17:21:41Z" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.843212 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.843262 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.843274 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.843292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.843308 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.946079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.946122 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.946135 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.946153 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:55 crc kubenswrapper[4761]: I1125 22:57:55.946165 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:55Z","lastTransitionTime":"2025-11-25T22:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.010321 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.010360 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:56 crc kubenswrapper[4761]: E1125 22:57:56.010450 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:56 crc kubenswrapper[4761]: E1125 22:57:56.010570 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.010613 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:56 crc kubenswrapper[4761]: E1125 22:57:56.010687 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.048899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.048962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.048978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.049008 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.049026 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.151440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.151495 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.151504 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.151521 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.151531 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.253832 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.253879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.253892 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.253911 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.253925 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.355245 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.355284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.355298 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.355311 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.355321 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.456885 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.456927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.456946 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.457031 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.457048 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.559832 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.559887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.559903 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.559938 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.559954 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.663069 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.663197 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.663254 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.663291 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.663873 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.766275 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.766315 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.766325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.766340 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.766353 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.868465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.868498 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.868509 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.868525 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.868536 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.970827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.970874 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.970890 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.970913 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:56 crc kubenswrapper[4761]: I1125 22:57:56.970931 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:56Z","lastTransitionTime":"2025-11-25T22:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.010363 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:57 crc kubenswrapper[4761]: E1125 22:57:57.010515 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.073475 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.073544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.073562 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.073590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.073609 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.176450 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.176706 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.176786 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.176804 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.176816 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.279005 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.279060 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.279068 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.279088 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.279097 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.382052 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.382125 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.382146 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.382176 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.382201 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.483858 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.483919 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.483936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.483960 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.483980 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.586415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.586812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.586958 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.587083 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.587205 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.689770 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.689818 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.689830 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.689846 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.689856 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.792496 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.792558 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.792575 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.792598 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.792615 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.894375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.894427 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.894445 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.894469 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.894485 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.997255 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.997510 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.997590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.997676 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:57 crc kubenswrapper[4761]: I1125 22:57:57.998136 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:57Z","lastTransitionTime":"2025-11-25T22:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.010590 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.010628 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:57:58 crc kubenswrapper[4761]: E1125 22:57:58.010747 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:57:58 crc kubenswrapper[4761]: E1125 22:57:58.010852 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.011008 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:57:58 crc kubenswrapper[4761]: E1125 22:57:58.011211 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.100193 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.100254 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.100267 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.100282 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.100295 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.202593 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.202673 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.202727 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.202749 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.202795 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.304871 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.305024 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.305083 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.305148 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.305210 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.407331 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.407441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.407548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.407613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.407673 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.509530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.509573 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.509585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.509602 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.509620 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.612548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.612597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.612613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.612636 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.612652 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.714972 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.715017 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.715025 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.715039 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.715050 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.817541 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.817590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.817605 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.817623 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.817636 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.919812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.919864 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.919879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.919901 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:58 crc kubenswrapper[4761]: I1125 22:57:58.919917 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:58Z","lastTransitionTime":"2025-11-25T22:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.011177 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:57:59 crc kubenswrapper[4761]: E1125 22:57:59.011407 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.022043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.022102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.022119 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.022140 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.022157 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.125140 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.125212 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.125235 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.125265 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.125286 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.228637 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.228679 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.228686 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.228720 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.228729 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.331816 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.331894 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.331916 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.331945 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.331965 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.434422 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.434482 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.434500 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.434523 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.434541 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.537532 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.537606 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.537629 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.537659 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.537678 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.640784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.640846 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.640862 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.640885 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.640903 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.743313 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.743359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.743375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.743397 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.743414 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.846801 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.846860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.846878 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.846903 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.846924 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.949522 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.949569 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.949584 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.949607 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:57:59 crc kubenswrapper[4761]: I1125 22:57:59.949624 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:57:59Z","lastTransitionTime":"2025-11-25T22:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.010898 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.010947 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.010913 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:00 crc kubenswrapper[4761]: E1125 22:58:00.011102 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:00 crc kubenswrapper[4761]: E1125 22:58:00.011229 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:00 crc kubenswrapper[4761]: E1125 22:58:00.011331 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.052021 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.052071 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.052087 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.052107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.052125 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.154532 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.154593 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.154610 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.154634 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.154652 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.257312 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.257395 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.257420 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.257453 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.257475 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.360462 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.360499 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.360507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.360520 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.360528 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.462897 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.462962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.462980 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.463003 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.463023 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.566067 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.566173 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.566195 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.566233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.566254 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.668761 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.668910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.668929 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.668956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.668973 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.771912 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.771980 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.772003 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.772033 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.772056 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.875366 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.875442 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.875461 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.875491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.875511 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.978432 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.978504 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.978522 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.978548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:00 crc kubenswrapper[4761]: I1125 22:58:00.978566 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:00Z","lastTransitionTime":"2025-11-25T22:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.011319 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:01 crc kubenswrapper[4761]: E1125 22:58:01.011477 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.081317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.081407 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.081425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.081447 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.081464 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.184036 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.184086 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.184103 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.184124 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.184141 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.287668 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.287769 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.287791 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.287821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.287838 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.390112 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.390199 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.390223 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.390256 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.390280 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.493921 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.494026 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.494076 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.494103 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.494121 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.598185 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.598265 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.598290 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.598326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.598349 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.701918 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.702020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.702073 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.702099 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.702116 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.805118 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.805207 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.805228 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.805281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.805299 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.908661 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.908849 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.908872 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.908895 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:01 crc kubenswrapper[4761]: I1125 22:58:01.908913 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:01Z","lastTransitionTime":"2025-11-25T22:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.010151 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.010175 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.010178 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.010327 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.010462 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.010636 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.012232 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.012326 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.012345 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.012371 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.012391 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.127541 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.127597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.127612 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.127639 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.127657 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.231389 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.231452 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.231470 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.231495 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.231512 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.334629 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.334677 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.334694 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.334744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.334762 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.438368 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.438436 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.438460 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.438491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.438515 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.541609 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.541661 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.541677 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.541728 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.541748 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.645234 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.645303 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.645319 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.645343 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.645369 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.748109 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.748183 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.748206 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.748233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.748250 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.851774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.851885 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.851903 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.851969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.851987 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.895205 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.895276 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.895298 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.895329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.895352 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.915990 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:02Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.921192 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.921269 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.921292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.921325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.921348 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.942927 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:02Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.949339 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.949602 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.949821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.950062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.950237 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:02 crc kubenswrapper[4761]: E1125 22:58:02.974900 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:02Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.980853 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.980918 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.980937 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.980960 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:02 crc kubenswrapper[4761]: I1125 22:58:02.980978 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:02Z","lastTransitionTime":"2025-11-25T22:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: E1125 22:58:03.003232 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:02Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.008345 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.008422 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.008443 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.008476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.008498 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.011145 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:03 crc kubenswrapper[4761]: E1125 22:58:03.011597 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.025835 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 22:58:03 crc kubenswrapper[4761]: E1125 22:58:03.031594 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:03Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:03 crc kubenswrapper[4761]: E1125 22:58:03.031795 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.033477 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.033547 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.033571 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.033603 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.033631 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.137392 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.137470 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.137496 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.137527 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.137545 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.241050 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.241113 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.241130 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.241152 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.241169 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.344223 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.344298 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.344352 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.344380 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.344397 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.447107 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.447180 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.447205 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.447235 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.447256 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.549965 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.550036 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.550053 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.550077 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.550096 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.652642 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.652719 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.652736 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.652759 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.652775 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.755822 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.755927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.755947 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.756015 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.756033 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.859405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.859464 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.859481 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.859505 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.859522 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.962899 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.962969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.962989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.963012 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:03 crc kubenswrapper[4761]: I1125 22:58:03.963029 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:03Z","lastTransitionTime":"2025-11-25T22:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.010750 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.010792 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.010767 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:04 crc kubenswrapper[4761]: E1125 22:58:04.010925 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:04 crc kubenswrapper[4761]: E1125 22:58:04.011071 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:04 crc kubenswrapper[4761]: E1125 22:58:04.011260 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.066238 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.066285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.066305 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.066332 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.066352 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.169576 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.169620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.169635 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.169656 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.169672 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.272467 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.272510 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.272527 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.272548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.272564 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.375353 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.375744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.376108 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.376447 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.376858 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.479989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.480316 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.480585 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.480793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.480976 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.583474 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.583530 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.583547 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.583570 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.583595 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.685975 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.686028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.686045 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.686079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.686094 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.788411 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.788466 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.788484 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.788507 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.788525 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.891501 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.891549 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.891566 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.891593 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.891624 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.994986 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.995053 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.995071 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.995100 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:04 crc kubenswrapper[4761]: I1125 22:58:04.995118 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:04Z","lastTransitionTime":"2025-11-25T22:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.010751 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:05 crc kubenswrapper[4761]: E1125 22:58:05.011119 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.012026 4761 scope.go:117] "RemoveContainer" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.033959 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.058659 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.080847 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098429 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098492 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098513 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098541 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098563 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.098998 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538dd8e9-85d4-43f3-aebb-e5925f5f85d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ea965844e67665a77c8ace18305a9ec605227d3b518ebc30d38454c627b216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.121240 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.136427 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.158424 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.174189 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.190430 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.200644 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.200800 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.200830 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.200856 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.200877 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.213673 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.231863 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.247668 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.277988 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.298089 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.303889 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.303932 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.303957 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.303982 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.304206 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.312921 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.325897 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.346638 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.361163 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.376235 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.407956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.408146 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.408238 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.408341 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.408448 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.511341 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.511383 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.511396 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.511414 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.511427 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.584933 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/2.log" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.588954 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.589477 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.604091 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.614014 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.614042 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.614054 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.614069 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.614082 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.619509 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.646891 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.659624 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.670921 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538dd8e9-85d4-43f3-aebb-e5925f5f85d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ea965844e67665a77c8ace18305a9ec605227d3b518ebc30d38454c627b216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.683297 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.694257 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.710474 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.716369 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.716415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.716432 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.716455 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.716472 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.722440 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.732103 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.752054 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.762586 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.771674 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.784472 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.798515 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.810692 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.818685 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.818742 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.818754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.818768 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.818779 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.825227 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.847018 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:58:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.859545 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:05Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.920965 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.921018 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.921034 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.921054 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:05 crc kubenswrapper[4761]: I1125 22:58:05.921070 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:05Z","lastTransitionTime":"2025-11-25T22:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.010807 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.010852 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.011013 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:06 crc kubenswrapper[4761]: E1125 22:58:06.011133 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:06 crc kubenswrapper[4761]: E1125 22:58:06.011282 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:06 crc kubenswrapper[4761]: E1125 22:58:06.011514 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.023977 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.024028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.024046 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.024072 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.024091 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.127041 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.127126 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.127146 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.127175 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.127201 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.230657 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.230763 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.230782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.230807 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.230824 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.333810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.333866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.333885 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.333910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.333931 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.436896 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.437249 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.437468 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.437646 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.437918 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.540879 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.540935 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.540956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.540984 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.541009 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.644177 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.644243 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.644261 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.644288 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.644308 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.747953 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.748024 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.748042 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.748074 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.748097 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.852531 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.852594 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.852612 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.852641 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.852660 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.955544 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.955590 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.955599 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.955617 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:06 crc kubenswrapper[4761]: I1125 22:58:06.955626 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:06Z","lastTransitionTime":"2025-11-25T22:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.010753 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:07 crc kubenswrapper[4761]: E1125 22:58:07.010971 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.058635 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.058727 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.058747 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.058776 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.058794 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.162182 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.162261 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.162282 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.162308 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.162325 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.265521 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.265620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.265640 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.265663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.265681 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.369144 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.369237 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.369262 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.369295 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.369319 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.471760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.471817 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.471836 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.471859 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.471877 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.574381 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.574440 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.574457 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.574484 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.574503 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.598397 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/3.log" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.599510 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/2.log" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.604054 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" exitCode=1 Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.604144 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.604232 4761 scope.go:117] "RemoveContainer" containerID="2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.605337 4761 scope.go:117] "RemoveContainer" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" Nov 25 22:58:07 crc kubenswrapper[4761]: E1125 22:58:07.605669 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.628025 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.650093 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.670890 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.677754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.677794 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.677805 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.677884 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.677897 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.688184 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.716868 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:58:06Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:140\\\\nI1125 22:58:05.948076 6778 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 22:58:05.948191 6778 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 22:58:05.948269 6778 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 22:58:05.958102 6778 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:58:05.958164 6778 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:58:05.958260 6778 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:58:05.958301 6778 factory.go:656] Stopping watch factory\\\\nI1125 22:58:05.958339 6778 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:58:05.990089 6778 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 22:58:05.990129 6778 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 22:58:05.990190 6778 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:58:05.990222 6778 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:58:05.990379 6778 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:58:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.733894 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.755666 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.776040 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.780755 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.780816 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.780834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.780858 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.780876 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.800778 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.818764 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.834646 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538dd8e9-85d4-43f3-aebb-e5925f5f85d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ea965844e67665a77c8ace18305a9ec605227d3b518ebc30d38454c627b216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.853826 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.872214 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.884301 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.884364 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.884388 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.884419 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.884441 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.895538 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.916835 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.933040 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.955902 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.969041 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.981116 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:07Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.987927 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.987953 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.987961 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.987973 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:07 crc kubenswrapper[4761]: I1125 22:58:07.987984 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:07Z","lastTransitionTime":"2025-11-25T22:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.011088 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:08 crc kubenswrapper[4761]: E1125 22:58:08.011221 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.011296 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:08 crc kubenswrapper[4761]: E1125 22:58:08.011351 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.011400 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:08 crc kubenswrapper[4761]: E1125 22:58:08.011457 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.090985 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.091043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.091058 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.091079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.091093 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.194550 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.194596 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.194613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.194633 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.194648 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.297238 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.297312 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.297327 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.297345 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.297357 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.400663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.400768 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.400785 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.400810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.400828 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.504285 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.504339 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.504354 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.504376 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.504393 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.607928 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.607992 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.608018 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.608047 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.608071 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.611522 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/3.log" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.711381 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.711447 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.711463 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.711488 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.711505 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.815380 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.815431 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.815449 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.815473 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.815491 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.918675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.918776 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.918792 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.918818 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:08 crc kubenswrapper[4761]: I1125 22:58:08.918837 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:08Z","lastTransitionTime":"2025-11-25T22:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.011196 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.012174 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.021129 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.021187 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.021210 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.021236 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.021258 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.124909 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.124973 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.124990 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.125014 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.125032 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.227690 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.227793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.227846 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.227868 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.227887 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.331001 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.331063 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.331080 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.331102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.331121 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.433866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.433925 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.433944 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.433967 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.433985 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.536598 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.536648 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.536739 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.536792 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.536814 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.639854 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.639897 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.639908 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.639928 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.639973 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.742795 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.742855 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.742872 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.742895 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.742913 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.784681 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.784916 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.784884289 +0000 UTC m=+149.518230164 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.852307 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.852772 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.852802 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.852845 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.852864 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.885893 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.886154 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886067 4761 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886280 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886505 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.886452519 +0000 UTC m=+149.619798384 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886575 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886620 4761 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.886782 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.886749536 +0000 UTC m=+149.620095411 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.886999 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887146 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887178 4761 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887197 4761 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887255 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.887239959 +0000 UTC m=+149.620585834 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.887454 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887554 4761 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: E1125 22:58:09.887614 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.887599377 +0000 UTC m=+149.620945242 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.956255 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.956318 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.956335 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.956361 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:09 crc kubenswrapper[4761]: I1125 22:58:09.956380 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:09Z","lastTransitionTime":"2025-11-25T22:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.010730 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.010803 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.010873 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:10 crc kubenswrapper[4761]: E1125 22:58:10.010938 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:10 crc kubenswrapper[4761]: E1125 22:58:10.011077 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:10 crc kubenswrapper[4761]: E1125 22:58:10.011370 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.059013 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.059308 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.059316 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.059329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.059338 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.162292 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.162346 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.162362 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.162386 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.162403 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.265222 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.265295 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.265320 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.265351 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.265373 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.368597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.368687 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.368744 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.368774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.368800 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.471941 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.472009 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.472032 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.472060 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.472081 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.575202 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.575247 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.575258 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.575321 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.575337 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.678013 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.678062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.678079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.678102 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.678119 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.781353 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.781405 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.781421 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.781445 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.781462 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.884565 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.884623 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.884644 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.884670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.884687 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.987852 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.987919 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.987942 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.987967 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:10 crc kubenswrapper[4761]: I1125 22:58:10.987986 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:10Z","lastTransitionTime":"2025-11-25T22:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.010814 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:11 crc kubenswrapper[4761]: E1125 22:58:11.011032 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.090620 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.090678 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.090726 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.090754 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.090776 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.193494 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.193560 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.193577 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.193601 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.193619 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.296614 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.296653 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.296663 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.296678 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.296688 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.398971 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.399021 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.399037 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.399059 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.399076 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.502277 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.502359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.502382 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.502408 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.502430 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.605385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.605456 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.605479 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.605506 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.605523 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.708505 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.708583 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.708602 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.708628 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.708647 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.812213 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.812339 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.812358 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.812407 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.812425 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.915158 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.915220 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.915241 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.915270 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:11 crc kubenswrapper[4761]: I1125 22:58:11.915292 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:11Z","lastTransitionTime":"2025-11-25T22:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.011184 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.011252 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.011185 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:12 crc kubenswrapper[4761]: E1125 22:58:12.011369 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:12 crc kubenswrapper[4761]: E1125 22:58:12.011610 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:12 crc kubenswrapper[4761]: E1125 22:58:12.011693 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.018637 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.018684 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.018729 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.018753 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.018772 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.121591 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.121635 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.121652 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.121675 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.121693 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.225519 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.225588 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.225606 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.225632 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.225650 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.328378 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.328431 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.328446 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.328465 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.328479 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.432170 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.432236 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.432296 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.432342 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.432367 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.536555 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.536613 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.536631 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.536655 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.536672 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.639964 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.640023 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.640039 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.640062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.640079 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.743172 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.743231 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.743248 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.743271 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.743287 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.846311 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.846374 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.846391 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.846415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.846433 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.948852 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.948910 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.948929 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.948953 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:12 crc kubenswrapper[4761]: I1125 22:58:12.948975 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:12Z","lastTransitionTime":"2025-11-25T22:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.010283 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.010468 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.051206 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.051278 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.051322 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.051345 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.051363 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.154517 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.154607 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.154627 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.154680 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.154731 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.221969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.222028 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.222051 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.222079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.222099 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.243124 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.248116 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.248173 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.248191 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.248219 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.248235 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.269473 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.274842 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.274913 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.274940 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.274965 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.274985 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.296084 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.301312 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.301375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.301396 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.301429 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.301452 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.322415 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.327387 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.327476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.327501 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.327536 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.327559 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.347971 4761 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T22:58:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a69c5537-0b60-4061-bda8-fe831ae1bd6e\\\",\\\"systemUUID\\\":\\\"262ec32c-2342-4d0c-b69e-19ea6b9dc450\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:13Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:13 crc kubenswrapper[4761]: E1125 22:58:13.348204 4761 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.351176 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.351241 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.351262 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.351296 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.351323 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.458635 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.458732 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.458765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.458792 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.458812 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.562437 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.562529 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.562548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.562571 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.562588 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.665611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.665682 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.665731 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.665760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.665778 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.769103 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.769148 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.769165 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.769195 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.769214 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.872866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.872942 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.872964 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.872996 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.873024 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.976234 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.976332 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.976349 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.976370 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:13 crc kubenswrapper[4761]: I1125 22:58:13.976385 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:13Z","lastTransitionTime":"2025-11-25T22:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.011377 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.011470 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.011390 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:14 crc kubenswrapper[4761]: E1125 22:58:14.011589 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:14 crc kubenswrapper[4761]: E1125 22:58:14.011939 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:14 crc kubenswrapper[4761]: E1125 22:58:14.012024 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.083441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.083518 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.083827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.083860 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.083887 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.187833 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.187951 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.187969 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.187993 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.188015 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.291141 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.291300 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.291330 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.291360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.291382 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.395020 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.395164 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.395183 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.395208 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.395267 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.498683 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.498787 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.498804 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.498828 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.498845 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.602444 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.602490 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.602502 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.602519 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.602531 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.706318 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.706750 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.706897 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.707086 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.707244 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.810805 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.810869 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.810891 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.810922 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.810941 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.914188 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.914267 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.914299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.914329 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:14 crc kubenswrapper[4761]: I1125 22:58:14.914353 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:14Z","lastTransitionTime":"2025-11-25T22:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.010580 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:15 crc kubenswrapper[4761]: E1125 22:58:15.010906 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.018825 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.018892 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.018912 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.018937 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.018957 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.045074 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c705fb5c076eca48283150b311428ca07c2b22c9a754f0778a187d08ee452ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:36Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1125 22:57:36.129473 6403 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1125 22:57:36.129561 6403 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 22:57:36.129620 6403 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 22:57:36.129840 6403 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:57:36.129866 6403 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 22:57:36.129907 6403 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 22:57:36.129935 6403 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:57:36.129967 6403 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130011 6403 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 22:57:36.130049 6403 factory.go:656] Stopping watch factory\\\\nI1125 22:57:36.130050 6403 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 22:57:36.130070 6403 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 22:57:36.130118 6403 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:57:36.130162 6403 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:57:36.130265 6403 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:58:06Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:140\\\\nI1125 22:58:05.948076 6778 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 22:58:05.948191 6778 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 22:58:05.948269 6778 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 22:58:05.958102 6778 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 22:58:05.958164 6778 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 22:58:05.958260 6778 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 22:58:05.958301 6778 factory.go:656] Stopping watch factory\\\\nI1125 22:58:05.958339 6778 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 22:58:05.990089 6778 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 22:58:05.990129 6778 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 22:58:05.990190 6778 ovnkube.go:599] Stopped ovnkube\\\\nI1125 22:58:05.990222 6778 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 22:58:05.990379 6778 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:58:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2lj29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qdk9h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.059914 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f76fd05-689a-44c0-9a1b-019930a01198\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdb7e87b3eb9709ecbf348a7335f5a9453bf74af75c94a511a4af122ec3d55c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a9e0f2645e701f866eaeced4de464a3dac3759935aef34d03c5717bcc3d8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n5cp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xppl7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.074446 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e42e251-564f-49d5-8a98-b13ee4ef6ad4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ce0a17df9813e51f922df65fdf6284bf11d36a3e064bdba46add383e4af8ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83360fb53a3367c8c9b0e65a45ae8a14300da4287123ef30c223bad03512b053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac00beebb1353a9d29dc646d6db448e0e8f8ef856a605e603aaac3482b8e4d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46772d07063bd0d3e21baa387edf4af318afb8ccb03137d2d63b6dd8a9b32715\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.092951 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.116165 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40da37c76667f338dd778f4adb3467932ea85c8df63c622ff14438188835cea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e033f7f12d551fcfa1f61db8f5da24f68a6bafe9d784c7b461202d6a138ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.121385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.121451 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.121476 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.121508 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.121532 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.133175 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6qz4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295217a2-b16b-4c48-ae61-00bfa20803fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03e0e457128370298bfc95995dec47f3eb005a6f62932421b206abd588a7fde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmrh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6qz4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.154473 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feda2a26-8d6f-4e04-b8b3-f606cd5408c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 22:56:58.449924 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 22:56:58.452405 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2374783610/tls.crt::/tmp/serving-cert-2374783610/tls.key\\\\\\\"\\\\nI1125 22:57:05.179257 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 22:57:05.184231 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 22:57:05.184256 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 22:57:05.184275 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 22:57:05.184282 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 22:57:05.191668 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 22:57:05.191709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 22:57:05.191731 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191742 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 22:57:05.191752 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 22:57:05.191758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 22:57:05.191764 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 22:57:05.191770 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 22:57:05.193475 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.176078 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8381730936d7b55e172f138756d310e6f0e8db7b942a2e454e58fdaf4cef9ff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.195334 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.218595 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7g447" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fda38150-e7bb-458c-8a8e-476ce0804735\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d321b4fae65770cd6fed0bde219c158049410754242c3066e7fab6633a93c1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0305daa0d7ec99c25d43cefcbcaf3089d0a06e7b952d160a933baa5953c12d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd02cd6b56658c9d53075d44b21483d0f05fbfe1a3fe3a5cc51406ee7efe2fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a830f03e50836534c681fdf0506061da22477f3c44afeba54ed245bfcc1b3a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f45b0df0d169faa99bda01bf5fd232d16538be7a9054441bd4bc97c5115f3a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19959790113c2c201f73ba3c996df29c5d87219ffcb3be71448b79f620398434\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0cb113628aacc7bcc1606ca69c2946fdba0a900a59fa365582e518ccf269395\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:57:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vvj9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7g447\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.223779 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.223834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.223851 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.223878 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.223895 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.241010 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fnqp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T22:57:54Z\\\",\\\"message\\\":\\\"2025-11-25T22:57:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6\\\\n2025-11-25T22:57:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5545fb6c-46ca-4f75-b016-c3074866d7b6 to /host/opt/cni/bin/\\\\n2025-11-25T22:57:09Z [verbose] multus-daemon started\\\\n2025-11-25T22:57:09Z [verbose] Readiness Indicator file check\\\\n2025-11-25T22:57:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lfl7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fnqp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.257898 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jvzpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9744f71-40a6-41d3-a8a3-6b1a16ea5740\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a253029b177c0a099c12dac49962581f25a200907923a0fbbba07f3ca46ed829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbj5z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jvzpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.278968 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a6ba0ea-5b97-47b2-a46a-ce8299dca67d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbd4db8eaa7e75480b3b2137879195ac5fea07290bfdd006dd344924b32baae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33bf164f9bfcda0ddd291ec763cfc5434f2045de8a09c08b057322d367094be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693c495348b76b5b7e2caa77c3e0d060cc23ed7531eb83f7f8650bd923836f8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.294787 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538dd8e9-85d4-43f3-aebb-e5925f5f85d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ea965844e67665a77c8ace18305a9ec605227d3b518ebc30d38454c627b216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d52cc0409963358fe0f8a631e51ae25f09c20048bfc6d77938104228b58e08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.309623 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.324563 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e27d0af-be2c-4fc2-93f7-2f5294537b9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a565becd9ea6fe7265712c3e2e3a7ef710102198e761cf335e058c7087195901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgsnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4lp9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.327799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.327889 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.327913 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.327945 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.327968 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.357639 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bbe600f8-9ad7-4aad-8a1e-da86319baacf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:56:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb9d01c82c12f075b09bdf597345a28ffd588dc073d2f90be085b73a67c54eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5cec269b067698ba15be4dca3ffdb92f6cae2870027856e29d2c547b3e62665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de0130b1429bff8ceb2fd3491da1786223d8d20cfe1501f7503b2c194083d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c1656a527614afd88d59f79c623497f2e3f5a98f176f12d5e19b4b1fce7ad1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db3d77207dfb5c101e6594b9ab38141fdb5caaa9248194afe6e3e09602333dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9f4d28ea16d366926c8da54150a860c09f371a74f8bc7dff92eb07c89a9785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd721ce741477f94e8696b3b0c1c88ac2d6a084dec9cf36d5c6a8e0c2d6b0bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d63b44533109301ed154c7dc46aafd17ff865fb3413fea6a74ee6ad2e30dd70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T22:56:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T22:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:56:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.376757 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d5c23a6dfe91e97fd832dccd77a4139156c8318b756834f1564016b9093b6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T22:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.397420 4761 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T22:57:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bl798\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T22:57:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhpwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T22:58:15Z is after 2025-08-24T17:21:41Z" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.431317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.431379 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.431396 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.431420 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.431440 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.533948 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.534010 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.534027 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.534052 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.534070 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.636760 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.637139 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.637288 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.637439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.637737 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.740302 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.740362 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.740379 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.740403 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.740421 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.843258 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.843309 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.843325 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.843347 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.843364 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.948664 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.948774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.948796 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.948821 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:15 crc kubenswrapper[4761]: I1125 22:58:15.948845 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:15Z","lastTransitionTime":"2025-11-25T22:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.010569 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.010598 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.010636 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:16 crc kubenswrapper[4761]: E1125 22:58:16.010827 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:16 crc kubenswrapper[4761]: E1125 22:58:16.010973 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:16 crc kubenswrapper[4761]: E1125 22:58:16.011060 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.051959 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.052021 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.052044 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.052074 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.052096 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.155415 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.155489 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.155509 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.155531 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.155551 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.259359 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.259434 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.259456 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.259510 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.259528 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.362679 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.362795 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.362817 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.362847 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.362872 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.466532 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.466600 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.466616 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.466640 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.466658 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.569471 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.569521 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.569533 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.569577 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.569591 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.672273 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.672347 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.672364 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.672399 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.672423 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.776037 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.776132 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.776154 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.776184 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.776208 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.878670 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.878765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.878782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.878809 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.878828 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.982156 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.982219 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.982232 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.982250 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:16 crc kubenswrapper[4761]: I1125 22:58:16.982265 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:16Z","lastTransitionTime":"2025-11-25T22:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.011536 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:17 crc kubenswrapper[4761]: E1125 22:58:17.011769 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.085957 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.086017 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.086032 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.086055 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.086073 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.189334 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.189408 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.189430 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.189457 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.189475 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.293098 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.293143 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.293155 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.293170 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.293181 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.396353 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.396449 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.396462 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.396491 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.396513 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.499876 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.499946 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.499960 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.499983 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.499997 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.603625 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.603756 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.603780 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.603816 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.603840 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.714799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.714871 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.714887 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.714912 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.714934 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.819022 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.819111 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.819137 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.819174 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.819198 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.922425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.922502 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.922525 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.922556 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:17 crc kubenswrapper[4761]: I1125 22:58:17.922578 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:17Z","lastTransitionTime":"2025-11-25T22:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.010542 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.010591 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.010549 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:18 crc kubenswrapper[4761]: E1125 22:58:18.010756 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:18 crc kubenswrapper[4761]: E1125 22:58:18.010885 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:18 crc kubenswrapper[4761]: E1125 22:58:18.011071 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.025481 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.025552 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.025573 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.025597 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.025614 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.129281 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.129372 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.129395 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.129426 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.129450 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.233184 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.233248 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.233260 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.233284 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.233296 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.336911 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.336995 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.337014 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.337042 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.337062 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.441155 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.441216 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.441233 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.441260 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.441278 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.544765 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.544812 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.544826 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.544846 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.544860 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.650766 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.650827 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.650848 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.650875 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.650899 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.754561 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.754612 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.754628 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.754651 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.754669 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.856532 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.856567 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.856577 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.856592 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.856601 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.958799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.958834 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.958842 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.958854 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:18 crc kubenswrapper[4761]: I1125 22:58:18.958862 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:18Z","lastTransitionTime":"2025-11-25T22:58:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.010814 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:19 crc kubenswrapper[4761]: E1125 22:58:19.011012 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.062216 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.062293 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.062317 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.062348 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.062368 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.164884 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.164951 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.164974 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.165006 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.165028 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.268450 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.268509 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.268525 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.268548 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.268566 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.372154 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.372258 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.372280 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.372312 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.372335 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.476274 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.476357 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.476375 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.476886 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.476949 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.580642 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.580740 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.580758 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.580784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.580803 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.683978 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.684070 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.684099 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.684134 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.684158 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.787078 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.787144 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.787168 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.787196 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.787220 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.891248 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.891331 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.891368 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.891399 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.891421 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.994395 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.994506 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.994524 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.994553 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:19 crc kubenswrapper[4761]: I1125 22:58:19.994574 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:19Z","lastTransitionTime":"2025-11-25T22:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.011245 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.011330 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.011251 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:20 crc kubenswrapper[4761]: E1125 22:58:20.011502 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:20 crc kubenswrapper[4761]: E1125 22:58:20.011736 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:20 crc kubenswrapper[4761]: E1125 22:58:20.011886 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.098322 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.098399 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.098422 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.098452 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.098473 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.202302 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.202369 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.202394 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.202424 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.202446 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.305232 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.305315 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.305334 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.305360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.305382 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.410439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.410622 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.410643 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.411360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.411395 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.515043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.515100 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.515118 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.515141 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.515160 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.619199 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.619309 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.619328 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.619352 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.619369 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.722310 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.722367 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.722388 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.722416 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.722435 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.825533 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.825591 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.825607 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.825628 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.825645 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.928439 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.928487 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.928502 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.928519 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:20 crc kubenswrapper[4761]: I1125 22:58:20.928531 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:20Z","lastTransitionTime":"2025-11-25T22:58:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.010650 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:21 crc kubenswrapper[4761]: E1125 22:58:21.010800 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.030576 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.030611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.030619 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.030632 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.030641 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.133454 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.133515 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.133533 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.133555 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.133572 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.237130 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.237203 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.237228 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.237256 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.237278 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.346215 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.346277 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.346299 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.346327 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.346349 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.449810 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.449870 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.449889 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.449912 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.449927 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.553057 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.553108 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.553126 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.553157 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.553180 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.656383 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.656425 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.656441 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.656458 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.656469 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.759511 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.759570 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.759589 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.759611 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.759629 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.861712 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.861764 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.861774 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.861791 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.861802 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.964866 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.964939 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.964956 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.964979 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:21 crc kubenswrapper[4761]: I1125 22:58:21.964996 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:21Z","lastTransitionTime":"2025-11-25T22:58:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.010614 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.010744 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.010778 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:22 crc kubenswrapper[4761]: E1125 22:58:22.010899 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:22 crc kubenswrapper[4761]: E1125 22:58:22.011000 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:22 crc kubenswrapper[4761]: E1125 22:58:22.011093 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.067793 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.067957 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.067965 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.067977 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.067985 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.171687 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.171782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.171799 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.171822 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.171850 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.274680 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.274784 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.274814 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.274839 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.274854 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.377570 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.377782 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.377819 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.377847 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.377869 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.480471 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.480523 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.480539 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.480563 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.480587 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.584274 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.584341 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.584360 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.584383 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.584402 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.687290 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.687349 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.687365 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.687389 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.687408 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.791079 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.791158 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.791178 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.791202 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.791219 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.893965 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.894005 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.894018 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.894062 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.894073 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.997932 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.997987 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.998004 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.998032 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:22 crc kubenswrapper[4761]: I1125 22:58:22.998050 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:22Z","lastTransitionTime":"2025-11-25T22:58:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.018432 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:23 crc kubenswrapper[4761]: E1125 22:58:23.018664 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.019276 4761 scope.go:117] "RemoveContainer" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" Nov 25 22:58:23 crc kubenswrapper[4761]: E1125 22:58:23.019654 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.075745 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.075685119 podStartE2EDuration="1m17.075685119s" podCreationTimestamp="2025-11-25 22:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.075264369 +0000 UTC m=+98.808610234" watchObservedRunningTime="2025-11-25 22:58:23.075685119 +0000 UTC m=+98.809030994" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.100871 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.100936 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.100962 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.100989 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.101013 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.177673 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podStartSLOduration=78.177655919 podStartE2EDuration="1m18.177655919s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.157467624 +0000 UTC m=+98.890813499" watchObservedRunningTime="2025-11-25 22:58:23.177655919 +0000 UTC m=+98.911001774" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.177876 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-7g447" podStartSLOduration=78.177870574 podStartE2EDuration="1m18.177870574s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.177310811 +0000 UTC m=+98.910656676" watchObservedRunningTime="2025-11-25 22:58:23.177870574 +0000 UTC m=+98.911216419" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.203000 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.203043 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.203054 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.203071 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.203084 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.208670 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fnqp9" podStartSLOduration=78.208650859 podStartE2EDuration="1m18.208650859s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.196440969 +0000 UTC m=+98.929786834" watchObservedRunningTime="2025-11-25 22:58:23.208650859 +0000 UTC m=+98.941996704" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.225360 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.225339738 podStartE2EDuration="1m18.225339738s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.225043761 +0000 UTC m=+98.958389646" watchObservedRunningTime="2025-11-25 22:58:23.225339738 +0000 UTC m=+98.958685593" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.225854 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jvzpt" podStartSLOduration=78.22584739 podStartE2EDuration="1m18.22584739s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.20910166 +0000 UTC m=+98.942447515" watchObservedRunningTime="2025-11-25 22:58:23.22584739 +0000 UTC m=+98.959193235" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.245269 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.245249586 podStartE2EDuration="20.245249586s" podCreationTimestamp="2025-11-25 22:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.235324103 +0000 UTC m=+98.968669998" watchObservedRunningTime="2025-11-25 22:58:23.245249586 +0000 UTC m=+98.978595431" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.277190 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.277169069 podStartE2EDuration="1m16.277169069s" podCreationTimestamp="2025-11-25 22:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.275871887 +0000 UTC m=+99.009217762" watchObservedRunningTime="2025-11-25 22:58:23.277169069 +0000 UTC m=+99.010514914" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.306103 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.306152 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.306164 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.306182 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.306194 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.310445 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6qz4k" podStartSLOduration=78.310431165 podStartE2EDuration="1m18.310431165s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.310329762 +0000 UTC m=+99.043675607" watchObservedRunningTime="2025-11-25 22:58:23.310431165 +0000 UTC m=+99.043777000" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.355678 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xppl7" podStartSLOduration=78.355646253 podStartE2EDuration="1m18.355646253s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.355275924 +0000 UTC m=+99.088621789" watchObservedRunningTime="2025-11-25 22:58:23.355646253 +0000 UTC m=+99.088992158" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.371724 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.371681185 podStartE2EDuration="46.371681185s" podCreationTimestamp="2025-11-25 22:57:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:23.371395529 +0000 UTC m=+99.104741394" watchObservedRunningTime="2025-11-25 22:58:23.371681185 +0000 UTC m=+99.105027040" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.408316 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.408370 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.408385 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.408407 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.408424 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.511097 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.511143 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.511159 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.511181 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.511198 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.575038 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.575165 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.575189 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.575215 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.575233 4761 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T22:58:23Z","lastTransitionTime":"2025-11-25T22:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.652174 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v"] Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.652811 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.656544 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.657556 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.657829 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.659338 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.743268 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.743345 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231c462-931c-4350-9b22-0b5e8e467fd2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.743484 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.743520 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4231c462-931c-4350-9b22-0b5e8e467fd2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.743567 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4231c462-931c-4350-9b22-0b5e8e467fd2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.844603 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.845100 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231c462-931c-4350-9b22-0b5e8e467fd2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.845347 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4231c462-931c-4350-9b22-0b5e8e467fd2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.844742 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.845532 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.845735 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4231c462-931c-4350-9b22-0b5e8e467fd2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.845921 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4231c462-931c-4350-9b22-0b5e8e467fd2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.847394 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4231c462-931c-4350-9b22-0b5e8e467fd2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.854280 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231c462-931c-4350-9b22-0b5e8e467fd2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.873358 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4231c462-931c-4350-9b22-0b5e8e467fd2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9v\" (UID: \"4231c462-931c-4350-9b22-0b5e8e467fd2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:23 crc kubenswrapper[4761]: I1125 22:58:23.977679 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.010485 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.010493 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:24 crc kubenswrapper[4761]: E1125 22:58:24.010725 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:24 crc kubenswrapper[4761]: E1125 22:58:24.010784 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.010957 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:24 crc kubenswrapper[4761]: E1125 22:58:24.011211 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.679746 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" event={"ID":"4231c462-931c-4350-9b22-0b5e8e467fd2","Type":"ContainerStarted","Data":"b0a967ba218625a029144fff4be331da033756f30e6520a4a9921bcacbb48f6c"} Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.680162 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" event={"ID":"4231c462-931c-4350-9b22-0b5e8e467fd2","Type":"ContainerStarted","Data":"96a10a9590c63cb0fdabfb76ae69e4267897c29dd7bf181446cd3c47640dd400"} Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.701000 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9v" podStartSLOduration=79.700969276 podStartE2EDuration="1m19.700969276s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:24.699646524 +0000 UTC m=+100.432992399" watchObservedRunningTime="2025-11-25 22:58:24.700969276 +0000 UTC m=+100.434315141" Nov 25 22:58:24 crc kubenswrapper[4761]: I1125 22:58:24.957163 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:24 crc kubenswrapper[4761]: E1125 22:58:24.957366 4761 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:58:24 crc kubenswrapper[4761]: E1125 22:58:24.957451 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs podName:2d99b4cb-ae29-4198-a407-a1cfe211f7a3 nodeName:}" failed. No retries permitted until 2025-11-25 22:59:28.957429404 +0000 UTC m=+164.690775249 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs") pod "network-metrics-daemon-lhpwb" (UID: "2d99b4cb-ae29-4198-a407-a1cfe211f7a3") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 22:58:25 crc kubenswrapper[4761]: I1125 22:58:25.012948 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:25 crc kubenswrapper[4761]: E1125 22:58:25.013119 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:26 crc kubenswrapper[4761]: I1125 22:58:26.010624 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:26 crc kubenswrapper[4761]: I1125 22:58:26.010652 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:26 crc kubenswrapper[4761]: I1125 22:58:26.010675 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:26 crc kubenswrapper[4761]: E1125 22:58:26.010823 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:26 crc kubenswrapper[4761]: E1125 22:58:26.010972 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:26 crc kubenswrapper[4761]: E1125 22:58:26.011143 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:27 crc kubenswrapper[4761]: I1125 22:58:27.010929 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:27 crc kubenswrapper[4761]: E1125 22:58:27.011195 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:28 crc kubenswrapper[4761]: I1125 22:58:28.010679 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:28 crc kubenswrapper[4761]: I1125 22:58:28.010864 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:28 crc kubenswrapper[4761]: I1125 22:58:28.011185 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:28 crc kubenswrapper[4761]: E1125 22:58:28.011172 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:28 crc kubenswrapper[4761]: E1125 22:58:28.011386 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:28 crc kubenswrapper[4761]: E1125 22:58:28.011529 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:29 crc kubenswrapper[4761]: I1125 22:58:29.010532 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:29 crc kubenswrapper[4761]: E1125 22:58:29.010808 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:30 crc kubenswrapper[4761]: I1125 22:58:30.010937 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:30 crc kubenswrapper[4761]: I1125 22:58:30.011016 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:30 crc kubenswrapper[4761]: I1125 22:58:30.011171 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:30 crc kubenswrapper[4761]: E1125 22:58:30.011344 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:30 crc kubenswrapper[4761]: E1125 22:58:30.011496 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:30 crc kubenswrapper[4761]: E1125 22:58:30.011739 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:31 crc kubenswrapper[4761]: I1125 22:58:31.010798 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:31 crc kubenswrapper[4761]: E1125 22:58:31.010977 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:32 crc kubenswrapper[4761]: I1125 22:58:32.010621 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:32 crc kubenswrapper[4761]: I1125 22:58:32.010685 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:32 crc kubenswrapper[4761]: I1125 22:58:32.010744 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:32 crc kubenswrapper[4761]: E1125 22:58:32.010830 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:32 crc kubenswrapper[4761]: E1125 22:58:32.010974 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:32 crc kubenswrapper[4761]: E1125 22:58:32.011097 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:33 crc kubenswrapper[4761]: I1125 22:58:33.010906 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:33 crc kubenswrapper[4761]: E1125 22:58:33.011153 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:34 crc kubenswrapper[4761]: I1125 22:58:34.010481 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:34 crc kubenswrapper[4761]: I1125 22:58:34.010740 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:34 crc kubenswrapper[4761]: E1125 22:58:34.010774 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:34 crc kubenswrapper[4761]: I1125 22:58:34.010832 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:34 crc kubenswrapper[4761]: E1125 22:58:34.010965 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:34 crc kubenswrapper[4761]: E1125 22:58:34.011087 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:35 crc kubenswrapper[4761]: I1125 22:58:35.010663 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:35 crc kubenswrapper[4761]: E1125 22:58:35.012653 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:36 crc kubenswrapper[4761]: I1125 22:58:36.011189 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:36 crc kubenswrapper[4761]: I1125 22:58:36.011246 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:36 crc kubenswrapper[4761]: E1125 22:58:36.011354 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:36 crc kubenswrapper[4761]: I1125 22:58:36.011382 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:36 crc kubenswrapper[4761]: E1125 22:58:36.011594 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:36 crc kubenswrapper[4761]: E1125 22:58:36.011837 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:37 crc kubenswrapper[4761]: I1125 22:58:37.010602 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:37 crc kubenswrapper[4761]: E1125 22:58:37.010729 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:38 crc kubenswrapper[4761]: I1125 22:58:38.010441 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:38 crc kubenswrapper[4761]: I1125 22:58:38.010450 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:38 crc kubenswrapper[4761]: E1125 22:58:38.010768 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:38 crc kubenswrapper[4761]: I1125 22:58:38.011013 4761 scope.go:117] "RemoveContainer" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" Nov 25 22:58:38 crc kubenswrapper[4761]: E1125 22:58:38.010994 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:38 crc kubenswrapper[4761]: E1125 22:58:38.011134 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qdk9h_openshift-ovn-kubernetes(f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" Nov 25 22:58:38 crc kubenswrapper[4761]: I1125 22:58:38.012028 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:38 crc kubenswrapper[4761]: E1125 22:58:38.012288 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:39 crc kubenswrapper[4761]: I1125 22:58:39.010415 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:39 crc kubenswrapper[4761]: E1125 22:58:39.010844 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.010992 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.011116 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.011327 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:40 crc kubenswrapper[4761]: E1125 22:58:40.011480 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:40 crc kubenswrapper[4761]: E1125 22:58:40.011740 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:40 crc kubenswrapper[4761]: E1125 22:58:40.011847 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.754071 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/1.log" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.754846 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/0.log" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.754911 4761 generic.go:334] "Generic (PLEG): container finished" podID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" containerID="aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577" exitCode=1 Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.754966 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerDied","Data":"aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577"} Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.755059 4761 scope.go:117] "RemoveContainer" containerID="e1f066a2f8a4d3541e68b34c05e6ae200ec75e559e1ccf2c95e69c78f30ffc0f" Nov 25 22:58:40 crc kubenswrapper[4761]: I1125 22:58:40.755736 4761 scope.go:117] "RemoveContainer" containerID="aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577" Nov 25 22:58:40 crc kubenswrapper[4761]: E1125 22:58:40.756030 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-fnqp9_openshift-multus(5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6)\"" pod="openshift-multus/multus-fnqp9" podUID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" Nov 25 22:58:41 crc kubenswrapper[4761]: I1125 22:58:41.010864 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:41 crc kubenswrapper[4761]: E1125 22:58:41.011059 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:41 crc kubenswrapper[4761]: I1125 22:58:41.761037 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/1.log" Nov 25 22:58:42 crc kubenswrapper[4761]: I1125 22:58:42.010571 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:42 crc kubenswrapper[4761]: I1125 22:58:42.010600 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:42 crc kubenswrapper[4761]: E1125 22:58:42.010769 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:42 crc kubenswrapper[4761]: I1125 22:58:42.010879 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:42 crc kubenswrapper[4761]: E1125 22:58:42.011096 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:42 crc kubenswrapper[4761]: E1125 22:58:42.011765 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:43 crc kubenswrapper[4761]: I1125 22:58:43.010622 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:43 crc kubenswrapper[4761]: E1125 22:58:43.010895 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:44 crc kubenswrapper[4761]: I1125 22:58:44.010830 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:44 crc kubenswrapper[4761]: I1125 22:58:44.011002 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:44 crc kubenswrapper[4761]: I1125 22:58:44.010825 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:44 crc kubenswrapper[4761]: E1125 22:58:44.011232 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:44 crc kubenswrapper[4761]: E1125 22:58:44.011369 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:44 crc kubenswrapper[4761]: E1125 22:58:44.011586 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:45 crc kubenswrapper[4761]: I1125 22:58:45.010588 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:45 crc kubenswrapper[4761]: E1125 22:58:45.012427 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:45 crc kubenswrapper[4761]: E1125 22:58:45.023076 4761 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 22:58:45 crc kubenswrapper[4761]: E1125 22:58:45.109293 4761 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:58:46 crc kubenswrapper[4761]: I1125 22:58:46.010270 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:46 crc kubenswrapper[4761]: I1125 22:58:46.010287 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:46 crc kubenswrapper[4761]: I1125 22:58:46.010419 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:46 crc kubenswrapper[4761]: E1125 22:58:46.010644 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:46 crc kubenswrapper[4761]: E1125 22:58:46.010768 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:46 crc kubenswrapper[4761]: E1125 22:58:46.010939 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:47 crc kubenswrapper[4761]: I1125 22:58:47.011239 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:47 crc kubenswrapper[4761]: E1125 22:58:47.011441 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:48 crc kubenswrapper[4761]: I1125 22:58:48.011059 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:48 crc kubenswrapper[4761]: E1125 22:58:48.011310 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:48 crc kubenswrapper[4761]: I1125 22:58:48.011753 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:48 crc kubenswrapper[4761]: E1125 22:58:48.011915 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:48 crc kubenswrapper[4761]: I1125 22:58:48.012181 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:48 crc kubenswrapper[4761]: E1125 22:58:48.012347 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:49 crc kubenswrapper[4761]: I1125 22:58:49.010351 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:49 crc kubenswrapper[4761]: E1125 22:58:49.010584 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:50 crc kubenswrapper[4761]: I1125 22:58:50.010254 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:50 crc kubenswrapper[4761]: I1125 22:58:50.010266 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:50 crc kubenswrapper[4761]: I1125 22:58:50.010397 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:50 crc kubenswrapper[4761]: E1125 22:58:50.011573 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:50 crc kubenswrapper[4761]: E1125 22:58:50.011755 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:50 crc kubenswrapper[4761]: E1125 22:58:50.011361 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:50 crc kubenswrapper[4761]: E1125 22:58:50.111085 4761 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:58:51 crc kubenswrapper[4761]: I1125 22:58:51.011296 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:51 crc kubenswrapper[4761]: E1125 22:58:51.011546 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:52 crc kubenswrapper[4761]: I1125 22:58:52.011289 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:52 crc kubenswrapper[4761]: I1125 22:58:52.011345 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:52 crc kubenswrapper[4761]: I1125 22:58:52.011341 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:52 crc kubenswrapper[4761]: E1125 22:58:52.011511 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:52 crc kubenswrapper[4761]: E1125 22:58:52.011817 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:52 crc kubenswrapper[4761]: E1125 22:58:52.012025 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.010672 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:53 crc kubenswrapper[4761]: E1125 22:58:53.010890 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.011343 4761 scope.go:117] "RemoveContainer" containerID="aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.012248 4761 scope.go:117] "RemoveContainer" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.808271 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/3.log" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.811569 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerStarted","Data":"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d"} Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.811997 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.813644 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/1.log" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.813720 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerStarted","Data":"438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4"} Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.846995 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podStartSLOduration=108.846964636 podStartE2EDuration="1m48.846964636s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:58:53.841676496 +0000 UTC m=+129.575022351" watchObservedRunningTime="2025-11-25 22:58:53.846964636 +0000 UTC m=+129.580310511" Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.975590 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lhpwb"] Nov 25 22:58:53 crc kubenswrapper[4761]: I1125 22:58:53.975746 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:53 crc kubenswrapper[4761]: E1125 22:58:53.975860 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:54 crc kubenswrapper[4761]: I1125 22:58:54.010869 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:54 crc kubenswrapper[4761]: I1125 22:58:54.010940 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:54 crc kubenswrapper[4761]: E1125 22:58:54.010983 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:54 crc kubenswrapper[4761]: E1125 22:58:54.011074 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:55 crc kubenswrapper[4761]: I1125 22:58:55.011118 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:55 crc kubenswrapper[4761]: E1125 22:58:55.013061 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:55 crc kubenswrapper[4761]: E1125 22:58:55.112357 4761 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:58:56 crc kubenswrapper[4761]: I1125 22:58:56.010646 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:56 crc kubenswrapper[4761]: I1125 22:58:56.010747 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:56 crc kubenswrapper[4761]: E1125 22:58:56.010844 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:56 crc kubenswrapper[4761]: I1125 22:58:56.010907 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:56 crc kubenswrapper[4761]: E1125 22:58:56.011050 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:56 crc kubenswrapper[4761]: E1125 22:58:56.011170 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:57 crc kubenswrapper[4761]: I1125 22:58:57.011109 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:57 crc kubenswrapper[4761]: E1125 22:58:57.011259 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:58:58 crc kubenswrapper[4761]: I1125 22:58:58.011069 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:58:58 crc kubenswrapper[4761]: E1125 22:58:58.011741 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:58:58 crc kubenswrapper[4761]: I1125 22:58:58.011434 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:58:58 crc kubenswrapper[4761]: E1125 22:58:58.011892 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:58:58 crc kubenswrapper[4761]: I1125 22:58:58.011382 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:58:58 crc kubenswrapper[4761]: E1125 22:58:58.011984 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:58:59 crc kubenswrapper[4761]: I1125 22:58:59.011040 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:58:59 crc kubenswrapper[4761]: E1125 22:58:59.011269 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 22:59:00 crc kubenswrapper[4761]: I1125 22:59:00.010284 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:59:00 crc kubenswrapper[4761]: I1125 22:59:00.010293 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:00 crc kubenswrapper[4761]: E1125 22:59:00.011389 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 22:59:00 crc kubenswrapper[4761]: E1125 22:59:00.011824 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 22:59:00 crc kubenswrapper[4761]: I1125 22:59:00.012170 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:59:00 crc kubenswrapper[4761]: E1125 22:59:00.012530 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhpwb" podUID="2d99b4cb-ae29-4198-a407-a1cfe211f7a3" Nov 25 22:59:01 crc kubenswrapper[4761]: I1125 22:59:01.011049 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:01 crc kubenswrapper[4761]: I1125 22:59:01.014576 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 22:59:01 crc kubenswrapper[4761]: I1125 22:59:01.017158 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.010736 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.010754 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.010750 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.014851 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.014970 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.015106 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 22:59:02 crc kubenswrapper[4761]: I1125 22:59:02.018052 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.492092 4761 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.542673 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.543524 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.547943 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.548812 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.550125 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.550289 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.550379 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.551958 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.552848 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.553094 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.553794 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.554047 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pss5v"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.554757 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.556903 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.557659 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.560186 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.561270 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.561445 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sdxpw"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.562403 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hcjtn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.562530 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.562921 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.562997 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563073 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563134 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563142 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563182 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563254 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563323 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563425 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5n2zp"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563550 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563738 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563803 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.563999 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.564382 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.564968 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.565732 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.565912 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.567864 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n5snq"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.568033 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.568638 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.568893 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.569171 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.569621 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.569622 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.569830 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.569851 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.570210 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.575154 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.575913 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-g4wpz"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.576460 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.579748 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fkj42"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.584321 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.586797 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.589758 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-etcd-client\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.589862 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n66s\" (UniqueName: \"kubernetes.io/projected/2aa3535f-06c8-4e56-ab10-589f653682aa-kube-api-access-9n66s\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.589948 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29rbd\" (UniqueName: \"kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590025 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-encryption-config\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590102 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-serving-cert\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590237 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590301 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590354 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/731312c6-273d-4d78-ab50-6a4f82fd61ae-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590407 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-config\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590468 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqhj2\" (UniqueName: \"kubernetes.io/projected/57448368-466e-484d-a45e-70c2ea89cc98-kube-api-access-dqhj2\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590516 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590611 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590664 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2aa3535f-06c8-4e56-ab10-589f653682aa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590744 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-audit-policies\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590826 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9rht\" (UniqueName: \"kubernetes.io/projected/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-kube-api-access-s9rht\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590875 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57448368-466e-484d-a45e-70c2ea89cc98-audit-dir\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.590958 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.591054 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.591098 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731312c6-273d-4d78-ab50-6a4f82fd61ae-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.591153 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-images\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.591205 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcbct\" (UniqueName: \"kubernetes.io/projected/731312c6-273d-4d78-ab50-6a4f82fd61ae-kube-api-access-qcbct\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.608627 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.609129 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.618070 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.618184 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.618626 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.618905 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619228 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619378 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619474 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619491 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619637 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619641 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619775 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619792 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.619989 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620099 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620218 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620348 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620459 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620562 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620664 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620785 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.620885 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.621016 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.621134 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.621286 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.622113 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.622231 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.622328 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.622525 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.622647 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.623308 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.623523 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.623680 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.623847 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.623984 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624104 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624229 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624355 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624559 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624685 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.624886 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.625257 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.625386 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.625684 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.625839 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626093 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626175 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626606 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626807 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626895 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.626966 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627029 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627182 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627334 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627420 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627504 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627562 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627598 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627662 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627676 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627767 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.627869 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.628778 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.629321 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.635022 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.635204 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.635395 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.636132 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.637659 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.637852 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.638342 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.638887 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.639079 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.641195 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.642054 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.642340 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.642559 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.642776 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.644349 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.644507 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.644625 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.646653 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.646911 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.647361 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hs5zk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.647557 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.647688 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.649005 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.649923 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.682855 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.684500 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.685187 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.685976 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.686464 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.687627 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.688146 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.688583 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.689066 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.689592 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.689659 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.690425 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691345 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691897 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-audit-policies\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691926 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9rht\" (UniqueName: \"kubernetes.io/projected/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-kube-api-access-s9rht\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691951 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5e072e1-338c-40a2-b8a5-88147f590a28-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691974 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.691993 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57448368-466e-484d-a45e-70c2ea89cc98-audit-dir\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692009 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692024 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692040 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692098 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692124 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692139 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692163 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731312c6-273d-4d78-ab50-6a4f82fd61ae-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692179 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692195 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56k6p\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-kube-api-access-56k6p\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692211 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcbct\" (UniqueName: \"kubernetes.io/projected/731312c6-273d-4d78-ab50-6a4f82fd61ae-kube-api-access-qcbct\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692227 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-images\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692255 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692270 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692286 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5e072e1-338c-40a2-b8a5-88147f590a28-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692304 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-etcd-client\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692322 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692341 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692357 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n66s\" (UniqueName: \"kubernetes.io/projected/2aa3535f-06c8-4e56-ab10-589f653682aa-kube-api-access-9n66s\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692389 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29rbd\" (UniqueName: \"kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692409 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-encryption-config\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692441 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692460 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692477 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-serving-cert\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692493 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692510 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692528 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692543 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/731312c6-273d-4d78-ab50-6a4f82fd61ae-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692559 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-config\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692576 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqhj2\" (UniqueName: \"kubernetes.io/projected/57448368-466e-484d-a45e-70c2ea89cc98-kube-api-access-dqhj2\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692591 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqvvd\" (UniqueName: \"kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692606 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692621 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692681 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692738 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2aa3535f-06c8-4e56-ab10-589f653682aa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.692824 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57448368-466e-484d-a45e-70c2ea89cc98-audit-dir\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.693003 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-audit-policies\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.693610 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.693642 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5jsv6"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.694971 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.695510 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.695577 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/731312c6-273d-4d78-ab50-6a4f82fd61ae-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.695910 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-config\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.696330 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/57448368-466e-484d-a45e-70c2ea89cc98-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.697977 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.698189 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2aa3535f-06c8-4e56-ab10-589f653682aa-images\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.699709 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.701214 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.702955 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731312c6-273d-4d78-ab50-6a4f82fd61ae-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.703039 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-serving-cert\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.706130 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.706914 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-encryption-config\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.707295 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/57448368-466e-484d-a45e-70c2ea89cc98-etcd-client\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.708760 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2aa3535f-06c8-4e56-ab10-589f653682aa-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.712073 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.712887 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.713003 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.713163 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.717372 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.723210 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5n2zp"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.724551 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sdxpw"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.726751 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.728392 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pss5v"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.729455 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pr4rh"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.730116 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.730516 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.731281 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.732890 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.733154 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.733990 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-wlhcm"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.734216 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.735075 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.735272 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.736712 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.736510 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.737731 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g6rkl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.737879 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.740661 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jrd27"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.740818 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.743556 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.743915 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.745155 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.745206 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.745473 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.745675 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.746070 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.746168 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.746397 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.748931 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fkj42"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.749030 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.755656 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.757505 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.761223 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g4wpz"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.762905 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.765327 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.765851 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.767170 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.770617 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.773461 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.775201 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n5snq"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.776452 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.778063 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.781193 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.782470 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.783670 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pr4rh"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.785193 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hcjtn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.786564 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hs5zk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.787600 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.787865 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5jsv6"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.788937 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.789994 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.791153 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g6rkl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.792210 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793260 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-k8fs6"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793435 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793467 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqvvd\" (UniqueName: \"kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793491 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793525 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5e072e1-338c-40a2-b8a5-88147f590a28-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793548 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793566 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793581 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793600 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793615 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794336 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794370 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56k6p\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-kube-api-access-56k6p\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794395 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794413 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794428 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5e072e1-338c-40a2-b8a5-88147f590a28-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794141 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794444 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794524 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.793959 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794581 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.794607 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.795189 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.795272 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.795336 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.795823 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jrd27"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.796067 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5e072e1-338c-40a2-b8a5-88147f590a28-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.796289 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.796953 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.797920 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.800253 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.800283 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.800457 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.800460 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.800605 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.802576 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5e072e1-338c-40a2-b8a5-88147f590a28-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.803961 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.805503 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.807014 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.808567 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7jq7c"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.809334 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.810402 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.812071 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.812347 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.815419 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.816028 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.817660 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-k8fs6"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.819262 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.820804 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vxvl7"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.821750 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.822417 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vxvl7"] Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.825974 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.845544 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.865953 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.885499 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.905548 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.925460 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.945976 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.967626 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 22:59:04 crc kubenswrapper[4761]: I1125 22:59:04.985868 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.006313 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.025469 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.047081 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.065945 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.085911 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.106970 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.137646 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.186585 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.205527 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.226270 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.247094 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.277160 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.285894 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.306356 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.326390 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.346117 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.366128 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.386751 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.406907 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.426489 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.446368 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.467172 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.514553 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9rht\" (UniqueName: \"kubernetes.io/projected/a0a04c8a-fa54-4164-bf6e-a14143aecfe8-kube-api-access-s9rht\") pod \"cluster-samples-operator-665b6dd947-s7vz9\" (UID: \"a0a04c8a-fa54-4164-bf6e-a14143aecfe8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.526546 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.533325 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcbct\" (UniqueName: \"kubernetes.io/projected/731312c6-273d-4d78-ab50-6a4f82fd61ae-kube-api-access-qcbct\") pod \"openshift-apiserver-operator-796bbdcf4f-nlqjr\" (UID: \"731312c6-273d-4d78-ab50-6a4f82fd61ae\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.547087 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.586172 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.596788 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n66s\" (UniqueName: \"kubernetes.io/projected/2aa3535f-06c8-4e56-ab10-589f653682aa-kube-api-access-9n66s\") pod \"machine-api-operator-5694c8668f-pss5v\" (UID: \"2aa3535f-06c8-4e56-ab10-589f653682aa\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.607344 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29rbd\" (UniqueName: \"kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd\") pod \"route-controller-manager-6576b87f9c-l4nsv\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.635114 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.635924 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqhj2\" (UniqueName: \"kubernetes.io/projected/57448368-466e-484d-a45e-70c2ea89cc98-kube-api-access-dqhj2\") pod \"apiserver-7bbb656c7d-xzlxb\" (UID: \"57448368-466e-484d-a45e-70c2ea89cc98\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.646937 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.665826 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.686680 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.706023 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.725666 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.757616 4761 request.go:700] Waited for 1.024227804s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.761108 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.776524 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.786157 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.788209 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.805947 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.811763 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.827937 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.836300 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.845636 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.866126 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.869246 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.886061 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.907580 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.925891 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.925998 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9"] Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.946498 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.966302 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 22:59:05 crc kubenswrapper[4761]: I1125 22:59:05.985814 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.006679 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.025636 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.046227 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.065660 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.085624 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.105457 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.112570 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pss5v"] Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.126492 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.146121 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.167404 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.185618 4761 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.205352 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.218583 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr"] Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.229423 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.247564 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.261271 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.266302 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.285648 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.286415 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb"] Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.305545 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.326136 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.346194 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.365940 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.385795 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.406159 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.441480 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.445192 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.465626 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.486476 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.525073 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.542751 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqvvd\" (UniqueName: \"kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd\") pod \"oauth-openshift-558db77b4-pp97f\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.545979 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.582142 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56k6p\" (UniqueName: \"kubernetes.io/projected/d5e072e1-338c-40a2-b8a5-88147f590a28-kube-api-access-56k6p\") pod \"cluster-image-registry-operator-dc59b4c8b-pswwn\" (UID: \"d5e072e1-338c-40a2-b8a5-88147f590a28\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.587182 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.606849 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.625668 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.645805 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.666165 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.685621 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.705309 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.726660 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.745413 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.764004 4761 request.go:700] Waited for 1.615855178s due to client-side throttling, not priority and fairness, request: PATCH:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/pods/machine-approver-56656f9798-l4zv4/status Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.801748 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819468 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819518 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8tgz\" (UniqueName: \"kubernetes.io/projected/18dc5416-2a42-4663-87dc-68c866e00af2-kube-api-access-t8tgz\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819579 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hplc\" (UniqueName: \"kubernetes.io/projected/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-kube-api-access-8hplc\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819626 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5aca764-c58d-45a9-bb38-60ea71fa9987-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819668 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-serving-cert\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819734 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-encryption-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819787 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-serving-cert\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.819925 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-service-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820082 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw6cl\" (UniqueName: \"kubernetes.io/projected/2244cddd-5bc9-46a0-86fe-64fa10c99ef1-kube-api-access-lw6cl\") pod \"migrator-59844c95c7-v8ftk\" (UID: \"2244cddd-5bc9-46a0-86fe-64fa10c99ef1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820144 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-oauth-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820200 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-etcd-client\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820252 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-etcd-serving-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820334 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820395 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-config\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820494 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-service-ca\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820599 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqwbb\" (UniqueName: \"kubernetes.io/projected/a0bdf401-28fd-4ce7-911b-e0402b22750c-kube-api-access-pqwbb\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820645 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820752 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-config\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820804 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820916 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c19247-5350-4d1c-a14d-d3e260b26bdd-serving-cert\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.820972 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821018 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-console-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821093 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/18dc5416-2a42-4663-87dc-68c866e00af2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821145 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nll6m\" (UniqueName: \"kubernetes.io/projected/c5aca764-c58d-45a9-bb38-60ea71fa9987-kube-api-access-nll6m\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821197 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdtcw\" (UniqueName: \"kubernetes.io/projected/b72410eb-f8eb-40c5-ae89-ba009453384c-kube-api-access-rdtcw\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821241 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-metrics-tls\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821330 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gqm4\" (UniqueName: \"kubernetes.io/projected/e1c19247-5350-4d1c-a14d-d3e260b26bdd-kube-api-access-9gqm4\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821877 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821938 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.821979 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-client\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822122 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-serving-cert\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822194 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/322a42f4-10a0-4c03-b9a7-3b3c2debf265-config\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822252 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbnxc\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-kube-api-access-rbnxc\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822398 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4dj\" (UniqueName: \"kubernetes.io/projected/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-kube-api-access-vz4dj\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822462 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-machine-approver-tls\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822507 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-trusted-ca\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822549 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-trusted-ca-bundle\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822594 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822643 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822694 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822814 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-oauth-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822859 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822899 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822959 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlr8r\" (UniqueName: \"kubernetes.io/projected/7219fd35-3830-495d-ac04-9b65ad696259-kube-api-access-zlr8r\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.822999 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/322a42f4-10a0-4c03-b9a7-3b3c2debf265-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823102 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-config\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823150 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823189 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mspmk\" (UniqueName: \"kubernetes.io/projected/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-kube-api-access-mspmk\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823254 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfn49\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823285 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823323 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-audit-dir\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823355 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5aca764-c58d-45a9-bb38-60ea71fa9987-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823443 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823483 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18dc5416-2a42-4663-87dc-68c866e00af2-serving-cert\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823513 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/322a42f4-10a0-4c03-b9a7-3b3c2debf265-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823544 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06ae2ec0-c649-45b6-b762-6827b898a96b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823574 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06ae2ec0-c649-45b6-b762-6827b898a96b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823652 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823682 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-service-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: E1125 22:59:06.823752 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.323680568 +0000 UTC m=+143.057026503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823814 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-audit\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823911 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-auth-proxy-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.823964 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.824009 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-image-import-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.824058 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-node-pullsecrets\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.867533 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.889181 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" event={"ID":"2aa3535f-06c8-4e56-ab10-589f653682aa","Type":"ContainerStarted","Data":"f40190ab377711090843ec96828427e388be8430b309cf37fb32268a7c4544fb"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.889251 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" event={"ID":"2aa3535f-06c8-4e56-ab10-589f653682aa","Type":"ContainerStarted","Data":"a690bb99e2a5a6a65503a609db02032cd1a0c7dea1d3ec522ed7f924ba56a64d"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.889273 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" event={"ID":"2aa3535f-06c8-4e56-ab10-589f653682aa","Type":"ContainerStarted","Data":"332d41cd19df7e96a51b44c6ca8138dbff5128019d449e030752301600eccd81"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.894140 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" event={"ID":"a0a04c8a-fa54-4164-bf6e-a14143aecfe8","Type":"ContainerStarted","Data":"86e74a4e258b29293d34d7f7fc51390b754fc463d53f494eb6e94c21a02b1d3c"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.894193 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" event={"ID":"a0a04c8a-fa54-4164-bf6e-a14143aecfe8","Type":"ContainerStarted","Data":"2dd6d40bea91ca119db328b99ae3c764f96d656a0cd8b64cafb7f7a9ee93fa01"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.894224 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" event={"ID":"a0a04c8a-fa54-4164-bf6e-a14143aecfe8","Type":"ContainerStarted","Data":"8f515d7bee27f3bcf4f70ce6fc252cf39bda69e92192f06098fd78ab1cca02f5"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.920339 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" event={"ID":"b5471e18-fe28-4269-bbd1-7e93dbd45b22","Type":"ContainerStarted","Data":"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.920388 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" event={"ID":"b5471e18-fe28-4269-bbd1-7e93dbd45b22","Type":"ContainerStarted","Data":"9b3ddcd8dc9e61131de61b8d7e1bb6e9ba077491ac280ef856830c9012eadbb3"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.923103 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.925055 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:06 crc kubenswrapper[4761]: E1125 22:59:06.925228 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.42520503 +0000 UTC m=+143.158550875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927239 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/322a42f4-10a0-4c03-b9a7-3b3c2debf265-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927300 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927331 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927378 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-config\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927406 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-profile-collector-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927458 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927486 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mspmk\" (UniqueName: \"kubernetes.io/projected/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-kube-api-access-mspmk\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927514 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-mountpoint-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927560 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927591 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5aca764-c58d-45a9-bb38-60ea71fa9987-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927619 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-serving-cert\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927647 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927679 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz7wt\" (UniqueName: \"kubernetes.io/projected/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-kube-api-access-xz7wt\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927727 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpcr6\" (UniqueName: \"kubernetes.io/projected/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-kube-api-access-jpcr6\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927762 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18dc5416-2a42-4663-87dc-68c866e00af2-serving-cert\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927793 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/322a42f4-10a0-4c03-b9a7-3b3c2debf265-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927861 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06ae2ec0-c649-45b6-b762-6827b898a96b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927934 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927968 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-service-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.927998 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-audit\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928030 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvtf8\" (UniqueName: \"kubernetes.io/projected/55868319-fab2-4d45-8176-099029f717bb-kube-api-access-lvtf8\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928063 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclxk\" (UniqueName: \"kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928094 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-image-import-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928142 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-node-pullsecrets\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928188 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928218 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928247 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hplc\" (UniqueName: \"kubernetes.io/projected/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-kube-api-access-8hplc\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928285 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/592cb417-405d-4a24-bd0b-b16f550e2f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928316 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-serving-cert\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928345 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928393 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whqbg\" (UniqueName: \"kubernetes.io/projected/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-kube-api-access-whqbg\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928421 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mgg4\" (UniqueName: \"kubernetes.io/projected/311fcb57-e225-42e2-8ba8-f21e7b83eb20-kube-api-access-4mgg4\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928450 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-certs\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928498 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gql82\" (UniqueName: \"kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928525 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928555 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-etcd-client\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928583 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z6ss\" (UniqueName: \"kubernetes.io/projected/08893509-ce28-44f6-a1fe-de9bfdd93020-kube-api-access-6z6ss\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928648 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqwbb\" (UniqueName: \"kubernetes.io/projected/a0bdf401-28fd-4ce7-911b-e0402b22750c-kube-api-access-pqwbb\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928677 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-service-ca\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928743 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/311fcb57-e225-42e2-8ba8-f21e7b83eb20-config-volume\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928771 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928804 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928832 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/311fcb57-e225-42e2-8ba8-f21e7b83eb20-metrics-tls\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928863 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/18dc5416-2a42-4663-87dc-68c866e00af2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928911 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdtcw\" (UniqueName: \"kubernetes.io/projected/b72410eb-f8eb-40c5-ae89-ba009453384c-kube-api-access-rdtcw\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928941 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-metrics-tls\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.928970 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nll6m\" (UniqueName: \"kubernetes.io/projected/c5aca764-c58d-45a9-bb38-60ea71fa9987-kube-api-access-nll6m\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929002 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcv95\" (UniqueName: \"kubernetes.io/projected/f7fec219-569e-45e0-8cff-7ca1c78139ab-kube-api-access-fcv95\") pod \"downloads-7954f5f757-5jsv6\" (UID: \"f7fec219-569e-45e0-8cff-7ca1c78139ab\") " pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929034 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9rsb\" (UniqueName: \"kubernetes.io/projected/592cb417-405d-4a24-bd0b-b16f550e2f9f-kube-api-access-v9rsb\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929082 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929118 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-client\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929149 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220f8c14-3aae-461d-958b-9487d7e9a7d5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929179 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-stats-auth\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929230 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zwll\" (UniqueName: \"kubernetes.io/projected/31a9c460-6b17-44cb-82ee-61ca104baeca-kube-api-access-5zwll\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929289 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-webhook-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929321 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-trusted-ca\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929351 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-trusted-ca-bundle\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: E1125 22:59:06.929410 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.429397185 +0000 UTC m=+143.162743020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929454 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929666 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929723 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlr8r\" (UniqueName: \"kubernetes.io/projected/7219fd35-3830-495d-ac04-9b65ad696259-kube-api-access-zlr8r\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929749 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-plugins-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929772 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929797 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-default-certificate\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929824 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfn49\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929846 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/142e22c9-656c-4130-b1ed-6b41db3f6170-cert\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929921 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-audit-dir\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929970 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930541 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/220f8c14-3aae-461d-958b-9487d7e9a7d5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930613 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930658 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84m8\" (UniqueName: \"kubernetes.io/projected/142e22c9-656c-4130-b1ed-6b41db3f6170-kube-api-access-s84m8\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930689 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06ae2ec0-c649-45b6-b762-6827b898a96b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930727 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-images\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930750 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-metrics-certs\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930777 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-auth-proxy-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930857 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/019554bd-ec80-42eb-8439-fe453d8537cc-service-ca-bundle\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930885 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-proxy-tls\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930957 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.930984 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc69q\" (UniqueName: \"kubernetes.io/projected/2a2a3111-ccac-4848-ae78-138daab3a2c6-kube-api-access-kc69q\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931071 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-csi-data-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931099 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mq6p\" (UniqueName: \"kubernetes.io/projected/019554bd-ec80-42eb-8439-fe453d8537cc-kube-api-access-2mq6p\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931161 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8tgz\" (UniqueName: \"kubernetes.io/projected/18dc5416-2a42-4663-87dc-68c866e00af2-kube-api-access-t8tgz\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931200 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5aca764-c58d-45a9-bb38-60ea71fa9987-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931239 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ed043b5-92aa-460d-8ced-1ac775019c08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931265 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-encryption-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931287 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vbn\" (UniqueName: \"kubernetes.io/projected/427ade36-c2c8-4e77-8139-59f950c4368a-kube-api-access-f7vbn\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931310 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49xqb\" (UniqueName: \"kubernetes.io/projected/4de408c1-7855-4a50-893e-668acfa82b2d-kube-api-access-49xqb\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931422 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-serving-cert\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931462 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931489 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw6cl\" (UniqueName: \"kubernetes.io/projected/2244cddd-5bc9-46a0-86fe-64fa10c99ef1-kube-api-access-lw6cl\") pod \"migrator-59844c95c7-v8ftk\" (UID: \"2244cddd-5bc9-46a0-86fe-64fa10c99ef1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931512 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-oauth-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931536 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-service-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931563 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-etcd-serving-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931585 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-key\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931611 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931685 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-config\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931732 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-config\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931757 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931779 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931829 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-config\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931854 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931882 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c19247-5350-4d1c-a14d-d3e260b26bdd-serving-cert\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931903 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-console-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931926 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed043b5-92aa-460d-8ced-1ac775019c08-config\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931949 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ed043b5-92aa-460d-8ced-1ac775019c08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.931981 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stkz7\" (UniqueName: \"kubernetes.io/projected/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-kube-api-access-stkz7\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932003 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-cabundle\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932025 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-socket-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932047 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-srv-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932068 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-apiservice-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932114 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gqm4\" (UniqueName: \"kubernetes.io/projected/e1c19247-5350-4d1c-a14d-d3e260b26bdd-kube-api-access-9gqm4\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932138 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-registration-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932165 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrmbt\" (UniqueName: \"kubernetes.io/projected/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-kube-api-access-mrmbt\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932195 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932220 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932244 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-serving-cert\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932268 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932290 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/31a9c460-6b17-44cb-82ee-61ca104baeca-tmpfs\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932312 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r9jm\" (UniqueName: \"kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932334 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-srv-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932363 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/322a42f4-10a0-4c03-b9a7-3b3c2debf265-config\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932388 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbnxc\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-kube-api-access-rbnxc\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932411 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2lk6\" (UniqueName: \"kubernetes.io/projected/220f8c14-3aae-461d-958b-9487d7e9a7d5-kube-api-access-m2lk6\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932450 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4dj\" (UniqueName: \"kubernetes.io/projected/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-kube-api-access-vz4dj\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932472 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-machine-approver-tls\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932492 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55868319-fab2-4d45-8176-099029f717bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932516 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-node-bootstrap-token\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932544 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932569 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932596 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.932619 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-oauth-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.941965 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-service-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.942203 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-etcd-serving-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.943008 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.943042 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5aca764-c58d-45a9-bb38-60ea71fa9987-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.943185 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.943829 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-auth-proxy-config\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.944061 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-config\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.944864 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.945067 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.945109 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-audit-dir\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.945733 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-config\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.966971 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.967613 4761 generic.go:334] "Generic (PLEG): container finished" podID="57448368-466e-484d-a45e-70c2ea89cc98" containerID="3e1fa7d450f0e93d451e6375a2d1b11834497eeb1e238023eefed07e8ef037a4" exitCode=0 Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.967773 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" event={"ID":"57448368-466e-484d-a45e-70c2ea89cc98","Type":"ContainerDied","Data":"3e1fa7d450f0e93d451e6375a2d1b11834497eeb1e238023eefed07e8ef037a4"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.967830 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" event={"ID":"57448368-466e-484d-a45e-70c2ea89cc98","Type":"ContainerStarted","Data":"670dab243eec8241ccb8cc682e3afdcca7429a2093bd99e453e179ec0a97f277"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.967883 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06ae2ec0-c649-45b6-b762-6827b898a96b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.967929 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c19247-5350-4d1c-a14d-d3e260b26bdd-serving-cert\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.969054 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-oauth-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.969269 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-console-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.970260 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-service-ca\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.970599 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.971013 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-serving-cert\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.971656 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18dc5416-2a42-4663-87dc-68c866e00af2-serving-cert\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.947610 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-encryption-config\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.929150 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-config\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.975782 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.969293 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.977229 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-oauth-config\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.982433 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5aca764-c58d-45a9-bb38-60ea71fa9987-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.982587 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-serving-cert\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.985603 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/322a42f4-10a0-4c03-b9a7-3b3c2debf265-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.985925 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" event={"ID":"731312c6-273d-4d78-ab50-6a4f82fd61ae","Type":"ContainerStarted","Data":"12a201b8ad67b7c4f1ec1fccbcff2c125583488c2bf1e40d255c88d73c2222b9"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.985982 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" event={"ID":"731312c6-273d-4d78-ab50-6a4f82fd61ae","Type":"ContainerStarted","Data":"f5e507e53df5d7140da6a075173a72749434a15e93325094ba47868bf4b10932"} Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.986445 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-machine-approver-tls\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.987167 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.987452 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/18dc5416-2a42-4663-87dc-68c866e00af2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.994656 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-audit\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.994910 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/322a42f4-10a0-4c03-b9a7-3b3c2debf265-config\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995109 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1c19247-5350-4d1c-a14d-d3e260b26bdd-trusted-ca\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995173 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7219fd35-3830-495d-ac04-9b65ad696259-node-pullsecrets\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995267 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-service-ca\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995404 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995770 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-metrics-tls\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.995958 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b72410eb-f8eb-40c5-ae89-ba009453384c-trusted-ca-bundle\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.996131 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b72410eb-f8eb-40c5-ae89-ba009453384c-console-serving-cert\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.997365 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0bdf401-28fd-4ce7-911b-e0402b22750c-etcd-client\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.998048 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/322a42f4-10a0-4c03-b9a7-3b3c2debf265-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z682m\" (UID: \"322a42f4-10a0-4c03-b9a7-3b3c2debf265\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:06 crc kubenswrapper[4761]: I1125 22:59:06.999610 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7219fd35-3830-495d-ac04-9b65ad696259-image-import-ca\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.000182 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.000654 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-etcd-client\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.001351 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mspmk\" (UniqueName: \"kubernetes.io/projected/fdd39b14-d07b-4dc6-86fa-72dc62f2c51b-kube-api-access-mspmk\") pod \"dns-operator-744455d44c-sdxpw\" (UID: \"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b\") " pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.002468 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06ae2ec0-c649-45b6-b762-6827b898a96b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.003947 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.007681 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7219fd35-3830-495d-ac04-9b65ad696259-serving-cert\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.010013 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c8de43-25a6-4f18-a39d-f388ba89eaa5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2tw9k\" (UID: \"a9c8de43-25a6-4f18-a39d-f388ba89eaa5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.025660 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gqm4\" (UniqueName: \"kubernetes.io/projected/e1c19247-5350-4d1c-a14d-d3e260b26bdd-kube-api-access-9gqm4\") pod \"console-operator-58897d9998-5n2zp\" (UID: \"e1c19247-5350-4d1c-a14d-d3e260b26bdd\") " pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.034289 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.035205 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gql82\" (UniqueName: \"kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.035237 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.035268 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z6ss\" (UniqueName: \"kubernetes.io/projected/08893509-ce28-44f6-a1fe-de9bfdd93020-kube-api-access-6z6ss\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.035381 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.535350307 +0000 UTC m=+143.268696212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.035642 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/311fcb57-e225-42e2-8ba8-f21e7b83eb20-config-volume\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.035673 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.036481 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/311fcb57-e225-42e2-8ba8-f21e7b83eb20-metrics-tls\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.037290 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.037460 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/311fcb57-e225-42e2-8ba8-f21e7b83eb20-config-volume\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038043 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcv95\" (UniqueName: \"kubernetes.io/projected/f7fec219-569e-45e0-8cff-7ca1c78139ab-kube-api-access-fcv95\") pod \"downloads-7954f5f757-5jsv6\" (UID: \"f7fec219-569e-45e0-8cff-7ca1c78139ab\") " pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038074 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9rsb\" (UniqueName: \"kubernetes.io/projected/592cb417-405d-4a24-bd0b-b16f550e2f9f-kube-api-access-v9rsb\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038099 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038122 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220f8c14-3aae-461d-958b-9487d7e9a7d5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038144 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-stats-auth\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038167 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zwll\" (UniqueName: \"kubernetes.io/projected/31a9c460-6b17-44cb-82ee-61ca104baeca-kube-api-access-5zwll\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038189 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-webhook-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038227 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-plugins-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038248 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038269 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-default-certificate\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038296 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/142e22c9-656c-4130-b1ed-6b41db3f6170-cert\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038327 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/220f8c14-3aae-461d-958b-9487d7e9a7d5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038347 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038370 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84m8\" (UniqueName: \"kubernetes.io/projected/142e22c9-656c-4130-b1ed-6b41db3f6170-kube-api-access-s84m8\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038407 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-images\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038423 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-metrics-certs\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038441 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/019554bd-ec80-42eb-8439-fe453d8537cc-service-ca-bundle\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038466 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-proxy-tls\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038500 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc69q\" (UniqueName: \"kubernetes.io/projected/2a2a3111-ccac-4848-ae78-138daab3a2c6-kube-api-access-kc69q\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038528 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-csi-data-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038551 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mq6p\" (UniqueName: \"kubernetes.io/projected/019554bd-ec80-42eb-8439-fe453d8537cc-kube-api-access-2mq6p\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038585 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ed043b5-92aa-460d-8ced-1ac775019c08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038623 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vbn\" (UniqueName: \"kubernetes.io/projected/427ade36-c2c8-4e77-8139-59f950c4368a-kube-api-access-f7vbn\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038647 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49xqb\" (UniqueName: \"kubernetes.io/projected/4de408c1-7855-4a50-893e-668acfa82b2d-kube-api-access-49xqb\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038673 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038752 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-key\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038790 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-config\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038815 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038843 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed043b5-92aa-460d-8ced-1ac775019c08-config\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038863 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ed043b5-92aa-460d-8ced-1ac775019c08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038885 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stkz7\" (UniqueName: \"kubernetes.io/projected/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-kube-api-access-stkz7\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038907 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-cabundle\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038927 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-socket-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038947 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-srv-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.038967 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-apiservice-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039005 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-registration-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039028 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrmbt\" (UniqueName: \"kubernetes.io/projected/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-kube-api-access-mrmbt\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039081 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039104 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/31a9c460-6b17-44cb-82ee-61ca104baeca-tmpfs\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039124 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r9jm\" (UniqueName: \"kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039145 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-srv-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039175 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2lk6\" (UniqueName: \"kubernetes.io/projected/220f8c14-3aae-461d-958b-9487d7e9a7d5-kube-api-access-m2lk6\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039217 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55868319-fab2-4d45-8176-099029f717bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039239 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-node-bootstrap-token\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039283 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039287 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039305 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039339 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-profile-collector-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039380 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039401 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-mountpoint-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039436 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-serving-cert\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039455 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039491 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz7wt\" (UniqueName: \"kubernetes.io/projected/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-kube-api-access-xz7wt\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039511 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpcr6\" (UniqueName: \"kubernetes.io/projected/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-kube-api-access-jpcr6\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039538 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvtf8\" (UniqueName: \"kubernetes.io/projected/55868319-fab2-4d45-8176-099029f717bb-kube-api-access-lvtf8\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039539 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039559 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclxk\" (UniqueName: \"kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039628 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039669 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/592cb417-405d-4a24-bd0b-b16f550e2f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039720 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039780 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whqbg\" (UniqueName: \"kubernetes.io/projected/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-kube-api-access-whqbg\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039805 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mgg4\" (UniqueName: \"kubernetes.io/projected/311fcb57-e225-42e2-8ba8-f21e7b83eb20-kube-api-access-4mgg4\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039857 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-certs\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.039963 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/311fcb57-e225-42e2-8ba8-f21e7b83eb20-metrics-tls\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.040739 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.041357 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.041605 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-webhook-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.041954 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/31a9c460-6b17-44cb-82ee-61ca104baeca-tmpfs\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.042248 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-registration-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.042511 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.043472 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-plugins-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.043595 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.047912 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/220f8c14-3aae-461d-958b-9487d7e9a7d5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.048510 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55868319-fab2-4d45-8176-099029f717bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.052143 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.552127889 +0000 UTC m=+143.285473814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.052450 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-mountpoint-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.053364 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/592cb417-405d-4a24-bd0b-b16f550e2f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.053464 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/220f8c14-3aae-461d-958b-9487d7e9a7d5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.053834 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/142e22c9-656c-4130-b1ed-6b41db3f6170-cert\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.054367 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/592cb417-405d-4a24-bd0b-b16f550e2f9f-images\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.054578 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.055150 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.055683 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-stats-auth\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.056192 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31a9c460-6b17-44cb-82ee-61ca104baeca-apiservice-cert\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.056284 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-csi-data-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.056286 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-default-certificate\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.057310 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-certs\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.057915 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058054 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058195 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/019554bd-ec80-42eb-8439-fe453d8537cc-metrics-certs\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058331 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4de408c1-7855-4a50-893e-668acfa82b2d-node-bootstrap-token\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058759 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-cabundle\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058855 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/427ade36-c2c8-4e77-8139-59f950c4368a-srv-cert\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058932 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/019554bd-ec80-42eb-8439-fe453d8537cc-service-ca-bundle\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.058962 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-socket-dir\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.059594 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-config\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.061953 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-serving-cert\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.062081 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed043b5-92aa-460d-8ced-1ac775019c08-config\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.062206 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ed043b5-92aa-460d-8ced-1ac775019c08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.062894 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-proxy-tls\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.063030 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2a2a3111-ccac-4848-ae78-138daab3a2c6-signing-key\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.063503 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.064234 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.064501 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-profile-collector-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.064691 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.066185 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/08893509-ce28-44f6-a1fe-de9bfdd93020-srv-cert\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.072451 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlr8r\" (UniqueName: \"kubernetes.io/projected/7219fd35-3830-495d-ac04-9b65ad696259-kube-api-access-zlr8r\") pod \"apiserver-76f77b778f-n5snq\" (UID: \"7219fd35-3830-495d-ac04-9b65ad696259\") " pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.081957 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.082870 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfn49\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.101585 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8tgz\" (UniqueName: \"kubernetes.io/projected/18dc5416-2a42-4663-87dc-68c866e00af2-kube-api-access-t8tgz\") pod \"openshift-config-operator-7777fb866f-fkj42\" (UID: \"18dc5416-2a42-4663-87dc-68c866e00af2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.124292 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw6cl\" (UniqueName: \"kubernetes.io/projected/2244cddd-5bc9-46a0-86fe-64fa10c99ef1-kube-api-access-lw6cl\") pod \"migrator-59844c95c7-v8ftk\" (UID: \"2244cddd-5bc9-46a0-86fe-64fa10c99ef1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.126505 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.134797 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.138966 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqwbb\" (UniqueName: \"kubernetes.io/projected/a0bdf401-28fd-4ce7-911b-e0402b22750c-kube-api-access-pqwbb\") pod \"etcd-operator-b45778765-hcjtn\" (UID: \"a0bdf401-28fd-4ce7-911b-e0402b22750c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.140606 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.140793 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.640771206 +0000 UTC m=+143.374117091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.140926 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.141563 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.641552645 +0000 UTC m=+143.374898480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.149892 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.154035 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:07 crc kubenswrapper[4761]: W1125 22:59:07.158827 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5e072e1_338c_40a2_b8a5_88147f590a28.slice/crio-f56929c80bb868dfe8049792108183537a577868e0038dfc5fb100fcb7658f49 WatchSource:0}: Error finding container f56929c80bb868dfe8049792108183537a577868e0038dfc5fb100fcb7658f49: Status 404 returned error can't find the container with id f56929c80bb868dfe8049792108183537a577868e0038dfc5fb100fcb7658f49 Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.160262 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbnxc\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-kube-api-access-rbnxc\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.162002 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.179341 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4dj\" (UniqueName: \"kubernetes.io/projected/6157c65a-f99d-4958-85f3-ec2e9ea06b8e-kube-api-access-vz4dj\") pod \"authentication-operator-69f744f599-hs5zk\" (UID: \"6157c65a-f99d-4958-85f3-ec2e9ea06b8e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.191460 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.205311 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hplc\" (UniqueName: \"kubernetes.io/projected/33bd1f91-64e1-40d9-97e3-08ba2c52ea97-kube-api-access-8hplc\") pod \"machine-approver-56656f9798-l4zv4\" (UID: \"33bd1f91-64e1-40d9-97e3-08ba2c52ea97\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.220433 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.226500 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdtcw\" (UniqueName: \"kubernetes.io/projected/b72410eb-f8eb-40c5-ae89-ba009453384c-kube-api-access-rdtcw\") pod \"console-f9d7485db-g4wpz\" (UID: \"b72410eb-f8eb-40c5-ae89-ba009453384c\") " pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.237119 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.242289 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.243120 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.743104887 +0000 UTC m=+143.476450712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.244938 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.247519 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nll6m\" (UniqueName: \"kubernetes.io/projected/c5aca764-c58d-45a9-bb38-60ea71fa9987-kube-api-access-nll6m\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz5xb\" (UID: \"c5aca764-c58d-45a9-bb38-60ea71fa9987\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.251180 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.258689 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.283689 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06ae2ec0-c649-45b6-b762-6827b898a96b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tm84l\" (UID: \"06ae2ec0-c649-45b6-b762-6827b898a96b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.303319 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gql82\" (UniqueName: \"kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82\") pod \"marketplace-operator-79b997595-cqrtc\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.326645 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z6ss\" (UniqueName: \"kubernetes.io/projected/08893509-ce28-44f6-a1fe-de9bfdd93020-kube-api-access-6z6ss\") pod \"catalog-operator-68c6474976-d7lr4\" (UID: \"08893509-ce28-44f6-a1fe-de9bfdd93020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.338875 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclxk\" (UniqueName: \"kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk\") pod \"controller-manager-879f6c89f-9p7tc\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.345791 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.346158 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.846148066 +0000 UTC m=+143.579493901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.359854 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84m8\" (UniqueName: \"kubernetes.io/projected/142e22c9-656c-4130-b1ed-6b41db3f6170-kube-api-access-s84m8\") pod \"ingress-canary-k8fs6\" (UID: \"142e22c9-656c-4130-b1ed-6b41db3f6170\") " pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.383270 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcv95\" (UniqueName: \"kubernetes.io/projected/f7fec219-569e-45e0-8cff-7ca1c78139ab-kube-api-access-fcv95\") pod \"downloads-7954f5f757-5jsv6\" (UID: \"f7fec219-569e-45e0-8cff-7ca1c78139ab\") " pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.401612 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9rsb\" (UniqueName: \"kubernetes.io/projected/592cb417-405d-4a24-bd0b-b16f550e2f9f-kube-api-access-v9rsb\") pod \"machine-config-operator-74547568cd-j4dlq\" (UID: \"592cb417-405d-4a24-bd0b-b16f550e2f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.407036 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.419465 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whqbg\" (UniqueName: \"kubernetes.io/projected/9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58-kube-api-access-whqbg\") pod \"csi-hostpathplugin-jrd27\" (UID: \"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58\") " pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.422016 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sdxpw"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.433895 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.436059 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.442254 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.445799 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-k8fs6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.446401 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.446874 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:07.946861386 +0000 UTC m=+143.680207221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.464862 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrmbt\" (UniqueName: \"kubernetes.io/projected/1477d4f4-8b05-4eb9-a972-bcc2408a3c00-kube-api-access-mrmbt\") pod \"multus-admission-controller-857f4d67dd-pr4rh\" (UID: \"1477d4f4-8b05-4eb9-a972-bcc2408a3c00\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.469109 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zwll\" (UniqueName: \"kubernetes.io/projected/31a9c460-6b17-44cb-82ee-61ca104baeca-kube-api-access-5zwll\") pod \"packageserver-d55dfcdfc-66j2w\" (UID: \"31a9c460-6b17-44cb-82ee-61ca104baeca\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.478932 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2lk6\" (UniqueName: \"kubernetes.io/projected/220f8c14-3aae-461d-958b-9487d7e9a7d5-kube-api-access-m2lk6\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrfp\" (UID: \"220f8c14-3aae-461d-958b-9487d7e9a7d5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.479161 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.530967 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.533584 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r9jm\" (UniqueName: \"kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm\") pod \"collect-profiles-29401845-vbw4h\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.539004 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mgg4\" (UniqueName: \"kubernetes.io/projected/311fcb57-e225-42e2-8ba8-f21e7b83eb20-kube-api-access-4mgg4\") pod \"dns-default-vxvl7\" (UID: \"311fcb57-e225-42e2-8ba8-f21e7b83eb20\") " pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.545297 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49xqb\" (UniqueName: \"kubernetes.io/projected/4de408c1-7855-4a50-893e-668acfa82b2d-kube-api-access-49xqb\") pod \"machine-config-server-7jq7c\" (UID: \"4de408c1-7855-4a50-893e-668acfa82b2d\") " pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.550078 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.550414 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.050402897 +0000 UTC m=+143.783748732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.563845 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc69q\" (UniqueName: \"kubernetes.io/projected/2a2a3111-ccac-4848-ae78-138daab3a2c6-kube-api-access-kc69q\") pod \"service-ca-9c57cc56f-g6rkl\" (UID: \"2a2a3111-ccac-4848-ae78-138daab3a2c6\") " pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.575822 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.576393 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.595996 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.599286 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpcr6\" (UniqueName: \"kubernetes.io/projected/5f54a75a-fa05-4a82-b40f-eba6b9cc3af7-kube-api-access-jpcr6\") pod \"control-plane-machine-set-operator-78cbb6b69f-t87pc\" (UID: \"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.600616 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ed043b5-92aa-460d-8ced-1ac775019c08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nwpc5\" (UID: \"4ed043b5-92aa-460d-8ced-1ac775019c08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.602164 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.614949 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.617772 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.621928 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stkz7\" (UniqueName: \"kubernetes.io/projected/2a3b9d72-ca02-45b2-ad09-63ad014e0eed-kube-api-access-stkz7\") pod \"machine-config-controller-84d6567774-wqpjh\" (UID: \"2a3b9d72-ca02-45b2-ad09-63ad014e0eed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.644488 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.651208 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.651564 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.151550649 +0000 UTC m=+143.884896484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.654424 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvtf8\" (UniqueName: \"kubernetes.io/projected/55868319-fab2-4d45-8176-099029f717bb-kube-api-access-lvtf8\") pod \"package-server-manager-789f6589d5-5rncl\" (UID: \"55868319-fab2-4d45-8176-099029f717bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.663624 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.673099 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.673139 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hs5zk"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.675526 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz7wt\" (UniqueName: \"kubernetes.io/projected/d5da9ea9-b02e-4112-9f0c-910aa1bc2229-kube-api-access-xz7wt\") pod \"service-ca-operator-777779d784-9xzkl\" (UID: \"d5da9ea9-b02e-4112-9f0c-910aa1bc2229\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.688561 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vbn\" (UniqueName: \"kubernetes.io/projected/427ade36-c2c8-4e77-8139-59f950c4368a-kube-api-access-f7vbn\") pod \"olm-operator-6b444d44fb-gh7tn\" (UID: \"427ade36-c2c8-4e77-8139-59f950c4368a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.690859 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n5snq"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.692936 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.696550 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hcjtn"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.704139 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.710985 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mq6p\" (UniqueName: \"kubernetes.io/projected/019554bd-ec80-42eb-8439-fe453d8537cc-kube-api-access-2mq6p\") pod \"router-default-5444994796-wlhcm\" (UID: \"019554bd-ec80-42eb-8439-fe453d8537cc\") " pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.720621 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.738522 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.760851 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.761012 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jq7c" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.761771 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.762074 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.262060566 +0000 UTC m=+143.995406411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.790676 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5n2zp"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.805790 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.862770 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.863759 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.36372944 +0000 UTC m=+144.097075275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.867983 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.868676 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.368647924 +0000 UTC m=+144.101993759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.882867 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.888034 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.926625 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.935504 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.938237 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fkj42"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.949816 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.963440 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk"] Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.969104 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:07 crc kubenswrapper[4761]: E1125 22:59:07.969531 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.469511988 +0000 UTC m=+144.202857823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.995577 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" event={"ID":"a0bdf401-28fd-4ce7-911b-e0402b22750c","Type":"ContainerStarted","Data":"691e41c80bab23028ab6699cc5c5fa4aa3b02b3700328424fde7ebcae11d7f4c"} Nov 25 22:59:07 crc kubenswrapper[4761]: I1125 22:59:07.997356 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" event={"ID":"322a42f4-10a0-4c03-b9a7-3b3c2debf265","Type":"ContainerStarted","Data":"30864b645f7a54597a8627a670d5cad458d3a72439afec4bf017d9bb68bfe5d5"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:07.999634 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" event={"ID":"33bd1f91-64e1-40d9-97e3-08ba2c52ea97","Type":"ContainerStarted","Data":"0c0f105825005cb7e17a97211563fc64dc16d9fff4691a8dbe0635541eb8ec44"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.001222 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" event={"ID":"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b","Type":"ContainerStarted","Data":"b6a70a37c982d1ad0ba3957e0acc263147a574b3205c2343b1ab18cf6a6f1a0d"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.004132 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" event={"ID":"401d0f53-1c92-4d21-a8f8-8989bb51b0b9","Type":"ContainerStarted","Data":"1c8c2eb17156c0dae8c83aec1b981e65efe99c12caa2753d4f49c481589a353c"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.004178 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" event={"ID":"401d0f53-1c92-4d21-a8f8-8989bb51b0b9","Type":"ContainerStarted","Data":"3d025499eaa640c96788cfca94e09553894cfe64c95f966310a83b8fb5ae1e1f"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.004596 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.028806 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" event={"ID":"57448368-466e-484d-a45e-70c2ea89cc98","Type":"ContainerStarted","Data":"43087034a047c18dc95ca3a3e07e8ac0b90c8a6f1b5d0a284d879f766f3f1077"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.033366 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" event={"ID":"6157c65a-f99d-4958-85f3-ec2e9ea06b8e","Type":"ContainerStarted","Data":"795f87a91625a13b51d424ccc16bfba4d7c8f1909c2dd91f505ddd70dd76b050"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.037938 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" event={"ID":"d5e072e1-338c-40a2-b8a5-88147f590a28","Type":"ContainerStarted","Data":"c98e668aaaf358eed9a430694a223a941c3789374afcb0fd9a133b43a95249e3"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.037964 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" event={"ID":"d5e072e1-338c-40a2-b8a5-88147f590a28","Type":"ContainerStarted","Data":"f56929c80bb868dfe8049792108183537a577868e0038dfc5fb100fcb7658f49"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.042579 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" event={"ID":"c5aca764-c58d-45a9-bb38-60ea71fa9987","Type":"ContainerStarted","Data":"70fddb52044de12d3dd39739b3e45fe71959b8ad0ce4cc25f9ac813838b9f509"} Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.063801 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.070220 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.071136 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.57112033 +0000 UTC m=+144.304466165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.117973 4761 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pp97f container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.118010 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.173022 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.174448 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.674432746 +0000 UTC m=+144.407778581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.276348 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.276801 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.776767707 +0000 UTC m=+144.510113542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.339753 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-pss5v" podStartSLOduration=123.33973714 podStartE2EDuration="2m3.33973714s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:08.339072983 +0000 UTC m=+144.072418828" watchObservedRunningTime="2025-11-25 22:59:08.33973714 +0000 UTC m=+144.073082975" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.383266 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.383608 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.883594392 +0000 UTC m=+144.616940227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.383653 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s7vz9" podStartSLOduration=123.383636273 podStartE2EDuration="2m3.383636273s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:08.382868723 +0000 UTC m=+144.116214558" watchObservedRunningTime="2025-11-25 22:59:08.383636273 +0000 UTC m=+144.116982108" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.436680 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" podStartSLOduration=123.436663795 podStartE2EDuration="2m3.436663795s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:08.436249295 +0000 UTC m=+144.169595140" watchObservedRunningTime="2025-11-25 22:59:08.436663795 +0000 UTC m=+144.170009640" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.486756 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.487188 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:08.987172005 +0000 UTC m=+144.720517840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.588622 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.588805 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.088781917 +0000 UTC m=+144.822127752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.590210 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.590588 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.090573942 +0000 UTC m=+144.823919777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.622486 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pswwn" podStartSLOduration=123.622458783 podStartE2EDuration="2m3.622458783s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:08.622029523 +0000 UTC m=+144.355375358" watchObservedRunningTime="2025-11-25 22:59:08.622458783 +0000 UTC m=+144.355804618" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.693469 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.695844 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.195819107 +0000 UTC m=+144.929164942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.695971 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.696860 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.196851112 +0000 UTC m=+144.930196947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.777589 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" podStartSLOduration=123.777569621 podStartE2EDuration="2m3.777569621s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:08.775551239 +0000 UTC m=+144.508897084" watchObservedRunningTime="2025-11-25 22:59:08.777569621 +0000 UTC m=+144.510915456" Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.796653 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.797084 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.29706443 +0000 UTC m=+145.030410265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:08 crc kubenswrapper[4761]: I1125 22:59:08.897982 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:08 crc kubenswrapper[4761]: E1125 22:59:08.898367 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.398349175 +0000 UTC m=+145.131695030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.000489 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.000672 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.500651235 +0000 UTC m=+145.233997080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.001346 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.001867 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.501854665 +0000 UTC m=+145.235200500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.060013 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-k8fs6"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.061732 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.062376 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jq7c" event={"ID":"4de408c1-7855-4a50-893e-668acfa82b2d","Type":"ContainerStarted","Data":"6e845d199e7d81414c1087f309a6668fd2b6d315ae8ede84b872d864b15b542a"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.062402 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jq7c" event={"ID":"4de408c1-7855-4a50-893e-668acfa82b2d","Type":"ContainerStarted","Data":"95952f49e4bdb24a11024661e7becacdc686a5558c42b53c0f030a63822e8b5b"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.064478 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wlhcm" event={"ID":"019554bd-ec80-42eb-8439-fe453d8537cc","Type":"ContainerStarted","Data":"5fb744c46a795f5eb274b43ec7c96e433d666e9e4e25f105452727030cb15395"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.066623 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" event={"ID":"6157c65a-f99d-4958-85f3-ec2e9ea06b8e","Type":"ContainerStarted","Data":"1f67cc75debb645bd3f3fd998d26a150a5c295a47bfd052b88e9b2c982f98639"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.071960 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" event={"ID":"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b","Type":"ContainerStarted","Data":"d8f64e48fc88317d56bab21e6a46dd4864e48ff3c9377aa391444bcb31e99a2d"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.071992 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" event={"ID":"fdd39b14-d07b-4dc6-86fa-72dc62f2c51b","Type":"ContainerStarted","Data":"57f826db4ca6a952b5cadb52420039af735e0cfcbcfc5ed1d07a2ccd3b9a2e3f"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.074088 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" event={"ID":"7219fd35-3830-495d-ac04-9b65ad696259","Type":"ContainerStarted","Data":"142b78121c12fb967064b27181580af9a140ef6d368e15105e31a3b5d5980fda"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.074113 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" event={"ID":"7219fd35-3830-495d-ac04-9b65ad696259","Type":"ContainerStarted","Data":"9a9d2d2ba5b304c44c3a5fc0f9a65bf94565dfd734ecb1ea9a3b441c11d85007"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.083754 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" event={"ID":"322a42f4-10a0-4c03-b9a7-3b3c2debf265","Type":"ContainerStarted","Data":"7797f1d166f66247d9055f8ffb12c7e57a7ef7e1b49d943a8e3c2ee43217999e"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.100313 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" event={"ID":"18dc5416-2a42-4663-87dc-68c866e00af2","Type":"ContainerStarted","Data":"af9592e17a5778d5446768fea6010ee38a63b17a6386f9960159676467632aa7"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.100356 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" event={"ID":"18dc5416-2a42-4663-87dc-68c866e00af2","Type":"ContainerStarted","Data":"b30b745c112bafbc92eda7a5afcadceae5ca547508cbd8b5ad93063773752000"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.102012 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.102328 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.60231403 +0000 UTC m=+145.335659865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.104209 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" event={"ID":"a0bdf401-28fd-4ce7-911b-e0402b22750c","Type":"ContainerStarted","Data":"97b2078c7075e2582506cd1f379655a63608bf3f860f8a8089a5e949b4787129"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.108006 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" event={"ID":"e1c19247-5350-4d1c-a14d-d3e260b26bdd","Type":"ContainerStarted","Data":"c0c918471b973855b6a63a929177d510b6f22d123aff19c9febae703a4573797"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.108036 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" event={"ID":"e1c19247-5350-4d1c-a14d-d3e260b26bdd","Type":"ContainerStarted","Data":"b2e869a3aa9db68e517dbfe4ab677cf1cc50e07f6b955dc11bc5382bb7e33f77"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.108464 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:09 crc kubenswrapper[4761]: W1125 22:59:09.116317 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7a16660_730c_4eb8_bfca_6ee890409ba0.slice/crio-8dbf497d525f2e3d09575a2602425f639f493613bb23c1fedd1d4af93ed25408 WatchSource:0}: Error finding container 8dbf497d525f2e3d09575a2602425f639f493613bb23c1fedd1d4af93ed25408: Status 404 returned error can't find the container with id 8dbf497d525f2e3d09575a2602425f639f493613bb23c1fedd1d4af93ed25408 Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.116505 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" event={"ID":"c5aca764-c58d-45a9-bb38-60ea71fa9987","Type":"ContainerStarted","Data":"365ae05a82eb56e8b4477defea4f63e7b68a2e13aef23f1aadd388045d4d88e7"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.137929 4761 patch_prober.go:28] interesting pod/console-operator-58897d9998-5n2zp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.137986 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" podUID="e1c19247-5350-4d1c-a14d-d3e260b26bdd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.158990 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" event={"ID":"2244cddd-5bc9-46a0-86fe-64fa10c99ef1","Type":"ContainerStarted","Data":"6de3a1b130a8f48370f3ac2f6ad208f00db74a5a59a332755e02f28137e450f1"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.159675 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" event={"ID":"2244cddd-5bc9-46a0-86fe-64fa10c99ef1","Type":"ContainerStarted","Data":"8bf19b5ed42858aaac1a66f56551789b932b02c094b4074ac0a521ed78359bce"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.162076 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" event={"ID":"fd8e84c9-b40f-4e03-a91a-508fd653ee07","Type":"ContainerStarted","Data":"bb07540777f25ef9946b9ce7193b39ddd1c7b72297220e8e032ad69253b5be91"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.162311 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.167736 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" podStartSLOduration=124.167711663 podStartE2EDuration="2m4.167711663s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.159080305 +0000 UTC m=+144.892426170" watchObservedRunningTime="2025-11-25 22:59:09.167711663 +0000 UTC m=+144.901057498" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.185926 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" event={"ID":"33bd1f91-64e1-40d9-97e3-08ba2c52ea97","Type":"ContainerStarted","Data":"59f3a7675c2b77e1b1a94e7d3f23aa8888722d72e0e2d64b917ecbed08fae838"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.189774 4761 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cqrtc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.189837 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.195086 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" event={"ID":"a9c8de43-25a6-4f18-a39d-f388ba89eaa5","Type":"ContainerStarted","Data":"33097c016f052ebc81b1f394ccc1ac7fbeb7c24d72a9e2f83a64518d090c1332"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.195144 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" event={"ID":"a9c8de43-25a6-4f18-a39d-f388ba89eaa5","Type":"ContainerStarted","Data":"ffa55ab15fb8576e92be1689122b3d4227c205b376eea6357abc59c8d4bf1bdc"} Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.222418 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.223625 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.225986 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.725971766 +0000 UTC m=+145.459317601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.236915 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5jsv6"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.276322 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.279576 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.282160 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g4wpz"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.299899 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp"] Nov 25 22:59:09 crc kubenswrapper[4761]: W1125 22:59:09.318871 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb72410eb_f8eb_40c5_ae89_ba009453384c.slice/crio-4eafc7142dbe5e64e68b66f3d55aff858c1d1d0e8e427c8a3365ae362e55338b WatchSource:0}: Error finding container 4eafc7142dbe5e64e68b66f3d55aff858c1d1d0e8e427c8a3365ae362e55338b: Status 404 returned error can't find the container with id 4eafc7142dbe5e64e68b66f3d55aff858c1d1d0e8e427c8a3365ae362e55338b Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.324095 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.326208 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.826183815 +0000 UTC m=+145.559529690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: W1125 22:59:09.343355 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06ae2ec0_c649_45b6_b762_6827b898a96b.slice/crio-128ee33b0e267720f8823d0eda8c0c034b3cd94f159c9bfd0f6253ebddbb6b66 WatchSource:0}: Error finding container 128ee33b0e267720f8823d0eda8c0c034b3cd94f159c9bfd0f6253ebddbb6b66: Status 404 returned error can't find the container with id 128ee33b0e267720f8823d0eda8c0c034b3cd94f159c9bfd0f6253ebddbb6b66 Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.376506 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nlqjr" podStartSLOduration=124.376482658 podStartE2EDuration="2m4.376482658s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.369952274 +0000 UTC m=+145.103298109" watchObservedRunningTime="2025-11-25 22:59:09.376482658 +0000 UTC m=+145.109828493" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.427354 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.427684 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:09.927672814 +0000 UTC m=+145.661018649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.454095 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.473972 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.511738 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.530164 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.530455 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.030440047 +0000 UTC m=+145.763785882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.566685 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" podStartSLOduration=124.566671116 podStartE2EDuration="2m4.566671116s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.520804974 +0000 UTC m=+145.254150809" watchObservedRunningTime="2025-11-25 22:59:09.566671116 +0000 UTC m=+145.300016951" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.604160 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hcjtn" podStartSLOduration=124.604137598 podStartE2EDuration="2m4.604137598s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.604079296 +0000 UTC m=+145.337425141" watchObservedRunningTime="2025-11-25 22:59:09.604137598 +0000 UTC m=+145.337483433" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.618627 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.632482 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.632945 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.132933302 +0000 UTC m=+145.866279137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.636800 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vxvl7"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.644425 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.654928 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.667307 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.674997 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" podStartSLOduration=124.674975448 podStartE2EDuration="2m4.674975448s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.652558065 +0000 UTC m=+145.385903910" watchObservedRunningTime="2025-11-25 22:59:09.674975448 +0000 UTC m=+145.408321283" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.699946 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.704059 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g6rkl"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.706319 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.709028 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz5xb" podStartSLOduration=124.709013073 podStartE2EDuration="2m4.709013073s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.681153773 +0000 UTC m=+145.414499618" watchObservedRunningTime="2025-11-25 22:59:09.709013073 +0000 UTC m=+145.442358908" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.710065 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jrd27"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.713006 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pr4rh"] Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.733149 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.733497 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.233482628 +0000 UTC m=+145.966828473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.787159 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" podStartSLOduration=124.787141736 podStartE2EDuration="2m4.787141736s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.721417715 +0000 UTC m=+145.454763560" watchObservedRunningTime="2025-11-25 22:59:09.787141736 +0000 UTC m=+145.520487571" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.834479 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.834799 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.334788283 +0000 UTC m=+146.068134118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.851491 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7jq7c" podStartSLOduration=5.851478082 podStartE2EDuration="5.851478082s" podCreationTimestamp="2025-11-25 22:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.851016451 +0000 UTC m=+145.584362286" watchObservedRunningTime="2025-11-25 22:59:09.851478082 +0000 UTC m=+145.584823917" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.923523 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z682m" podStartSLOduration=124.923501402 podStartE2EDuration="2m4.923501402s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.885067836 +0000 UTC m=+145.618413671" watchObservedRunningTime="2025-11-25 22:59:09.923501402 +0000 UTC m=+145.656847237" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.935015 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:09 crc kubenswrapper[4761]: E1125 22:59:09.935462 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.435445262 +0000 UTC m=+146.168791097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.967512 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2tw9k" podStartSLOduration=124.967492217 podStartE2EDuration="2m4.967492217s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.932863007 +0000 UTC m=+145.666208852" watchObservedRunningTime="2025-11-25 22:59:09.967492217 +0000 UTC m=+145.700838052" Nov 25 22:59:09 crc kubenswrapper[4761]: I1125 22:59:09.968145 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" podStartSLOduration=124.968137423 podStartE2EDuration="2m4.968137423s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.966746878 +0000 UTC m=+145.700092733" watchObservedRunningTime="2025-11-25 22:59:09.968137423 +0000 UTC m=+145.701483258" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.001102 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-hs5zk" podStartSLOduration=125.001088972 podStartE2EDuration="2m5.001088972s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:09.999266275 +0000 UTC m=+145.732612120" watchObservedRunningTime="2025-11-25 22:59:10.001088972 +0000 UTC m=+145.734434807" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.034643 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-sdxpw" podStartSLOduration=125.034626414 podStartE2EDuration="2m5.034626414s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.034056579 +0000 UTC m=+145.767402424" watchObservedRunningTime="2025-11-25 22:59:10.034626414 +0000 UTC m=+145.767972249" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.036652 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.036991 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.536977203 +0000 UTC m=+146.270323038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.137414 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.137768 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.637740074 +0000 UTC m=+146.371085929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.137969 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.138346 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.6383343 +0000 UTC m=+146.371680135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.237139 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" event={"ID":"06ae2ec0-c649-45b6-b762-6827b898a96b","Type":"ContainerStarted","Data":"a35c8194c9a5703dd936d7363b98b1ced3977657b28c9f3e98e3ae9aef29515f"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.237188 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" event={"ID":"06ae2ec0-c649-45b6-b762-6827b898a96b","Type":"ContainerStarted","Data":"128ee33b0e267720f8823d0eda8c0c034b3cd94f159c9bfd0f6253ebddbb6b66"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.238828 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.239053 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.73902469 +0000 UTC m=+146.472370525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.239160 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.239607 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.739595454 +0000 UTC m=+146.472941369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.243134 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" event={"ID":"220f8c14-3aae-461d-958b-9487d7e9a7d5","Type":"ContainerStarted","Data":"940206af1f99e8e43c42b08ed2f5add36069d7ad91d7f99fa086050d2292f522"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.243181 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" event={"ID":"220f8c14-3aae-461d-958b-9487d7e9a7d5","Type":"ContainerStarted","Data":"5c5a3a6c1ab96b99e3b5dd168e2ba17a99f8fe14d94298ed752d14b1c179efdf"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.276079 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrfp" podStartSLOduration=125.27606224 podStartE2EDuration="2m5.27606224s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.272323716 +0000 UTC m=+146.005669551" watchObservedRunningTime="2025-11-25 22:59:10.27606224 +0000 UTC m=+146.009408075" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.277785 4761 generic.go:334] "Generic (PLEG): container finished" podID="18dc5416-2a42-4663-87dc-68c866e00af2" containerID="af9592e17a5778d5446768fea6010ee38a63b17a6386f9960159676467632aa7" exitCode=0 Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.278084 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" event={"ID":"18dc5416-2a42-4663-87dc-68c866e00af2","Type":"ContainerDied","Data":"af9592e17a5778d5446768fea6010ee38a63b17a6386f9960159676467632aa7"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.282194 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" event={"ID":"4ed043b5-92aa-460d-8ced-1ac775019c08","Type":"ContainerStarted","Data":"ee0ad387846be5124eeb14f0647ee837ac836dbce13abd7c436a4c08c89f118a"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.286191 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-k8fs6" event={"ID":"142e22c9-656c-4130-b1ed-6b41db3f6170","Type":"ContainerStarted","Data":"fa818d46f271ecfd32357ffeb8d42ea8d44a9038dd2488dfee36c784ba40928a"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.286297 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-k8fs6" event={"ID":"142e22c9-656c-4130-b1ed-6b41db3f6170","Type":"ContainerStarted","Data":"a64e8b6497fb0dc51acf94c0bbd9689aec272634b067e00e583e0bb9d9c3448d"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.297941 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" event={"ID":"d7a16660-730c-4eb8-bfca-6ee890409ba0","Type":"ContainerStarted","Data":"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.297983 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" event={"ID":"d7a16660-730c-4eb8-bfca-6ee890409ba0","Type":"ContainerStarted","Data":"8dbf497d525f2e3d09575a2602425f639f493613bb23c1fedd1d4af93ed25408"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.298322 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.308877 4761 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9p7tc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.308934 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.310289 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" event={"ID":"31a9c460-6b17-44cb-82ee-61ca104baeca","Type":"ContainerStarted","Data":"31a92d94f50a3d2e67dab582433d8d18936cdf1206bf05c685e8425866992005"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.310332 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" event={"ID":"31a9c460-6b17-44cb-82ee-61ca104baeca","Type":"ContainerStarted","Data":"b415cfe6608ccfa83fe56272d3eb5fad5e9ff8d44128328301394dba17ae850b"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.313018 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.322739 4761 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-66j2w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.322788 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" podUID="31a9c460-6b17-44cb-82ee-61ca104baeca" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.322910 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" event={"ID":"1477d4f4-8b05-4eb9-a972-bcc2408a3c00","Type":"ContainerStarted","Data":"84c1f8059a82367e624001d8f359ec82f81a59ce8ff838579e84355c6a3e0ed4"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.345558 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" event={"ID":"55868319-fab2-4d45-8176-099029f717bb","Type":"ContainerStarted","Data":"689e10a523eb0bedd8819d3ea5f2431b2d4bfb82c0e6eb743dfb372cef9343b0"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.345607 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" event={"ID":"55868319-fab2-4d45-8176-099029f717bb","Type":"ContainerStarted","Data":"079088eaf1c57af9c550c71288cb954b728f0768a13f7b0ab034b0eba0afbe34"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.346290 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.346495 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.346691 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-k8fs6" podStartSLOduration=6.346672325 podStartE2EDuration="6.346672325s" podCreationTimestamp="2025-11-25 22:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.313242775 +0000 UTC m=+146.046588610" watchObservedRunningTime="2025-11-25 22:59:10.346672325 +0000 UTC m=+146.080018160" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.348350 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.848333036 +0000 UTC m=+146.581678871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.373581 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxvl7" event={"ID":"311fcb57-e225-42e2-8ba8-f21e7b83eb20","Type":"ContainerStarted","Data":"8e98219dea04cdd21f7b16b8018f58ed0b6e922f9c63f14b0570cbfdefb4fff0"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.379434 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g4wpz" event={"ID":"b72410eb-f8eb-40c5-ae89-ba009453384c","Type":"ContainerStarted","Data":"1d28bb61df3eb5797bae2034f0231bcf3581ba60f52a0102596a682f5976e0f0"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.379484 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g4wpz" event={"ID":"b72410eb-f8eb-40c5-ae89-ba009453384c","Type":"ContainerStarted","Data":"4eafc7142dbe5e64e68b66f3d55aff858c1d1d0e8e427c8a3365ae362e55338b"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.404826 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" podStartSLOduration=125.404810985 podStartE2EDuration="2m5.404810985s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.404242121 +0000 UTC m=+146.137587956" watchObservedRunningTime="2025-11-25 22:59:10.404810985 +0000 UTC m=+146.138156820" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.405163 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" podStartSLOduration=125.405159063 podStartE2EDuration="2m5.405159063s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.357595328 +0000 UTC m=+146.090941163" watchObservedRunningTime="2025-11-25 22:59:10.405159063 +0000 UTC m=+146.138504908" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.421807 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" event={"ID":"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3","Type":"ContainerStarted","Data":"2d492fa03dd4556dc52304c02765ee0c5c3fb367f85b35fe2ac0dcf12a26ff9f"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.442643 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5jsv6" event={"ID":"f7fec219-569e-45e0-8cff-7ca1c78139ab","Type":"ContainerStarted","Data":"d77cb3a8a1b0e959f2e6e9272a435400b99033acf77726d2d3305e5ca98e5beb"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.442722 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5jsv6" event={"ID":"f7fec219-569e-45e0-8cff-7ca1c78139ab","Type":"ContainerStarted","Data":"27fad5dedb427fd4ec768597e05d72ed97fce4c37ee7356c3de5575adc185c8c"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.443852 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" podStartSLOduration=125.443835685 podStartE2EDuration="2m5.443835685s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.442676826 +0000 UTC m=+146.176022671" watchObservedRunningTime="2025-11-25 22:59:10.443835685 +0000 UTC m=+146.177181520" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.443899 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.452098 4761 patch_prober.go:28] interesting pod/downloads-7954f5f757-5jsv6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.452353 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5jsv6" podUID="f7fec219-569e-45e0-8cff-7ca1c78139ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.452967 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.453243 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:10.953228032 +0000 UTC m=+146.686573877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.455379 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" event={"ID":"592cb417-405d-4a24-bd0b-b16f550e2f9f","Type":"ContainerStarted","Data":"bea2e3cb107e080a3c90aeee039f8d2e00b52209c1ed6fb6d7d62fa226c334ff"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.455427 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" event={"ID":"592cb417-405d-4a24-bd0b-b16f550e2f9f","Type":"ContainerStarted","Data":"bbadd977b9aa847b8ebcf5d4233e6662ca4aea6de42c3419245a7fc4347aaeac"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.460110 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" event={"ID":"08893509-ce28-44f6-a1fe-de9bfdd93020","Type":"ContainerStarted","Data":"8abea0ac347e7aed743d8cdd90d42aee1af229b8a218e60535fdd0a5b6885c24"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.460155 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" event={"ID":"08893509-ce28-44f6-a1fe-de9bfdd93020","Type":"ContainerStarted","Data":"f21f1d80be4e4710060c202058f382998c0f1c4c06944e9043d8b7e43bdec7b4"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.461008 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.468432 4761 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-d7lr4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.468471 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" podUID="08893509-ce28-44f6-a1fe-de9bfdd93020" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.473882 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" event={"ID":"427ade36-c2c8-4e77-8139-59f950c4368a","Type":"ContainerStarted","Data":"0d4080b236a4cf2a61227feec13837abb054ab7b154d614f9197264c490bf3c8"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.473918 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" event={"ID":"427ade36-c2c8-4e77-8139-59f950c4368a","Type":"ContainerStarted","Data":"0f82efd75d7b8d8c59387d712b70f268a662e31f25517f741639c88fbbd29a95"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.474400 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.476132 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-g4wpz" podStartSLOduration=125.476116816 podStartE2EDuration="2m5.476116816s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.473115851 +0000 UTC m=+146.206461696" watchObservedRunningTime="2025-11-25 22:59:10.476116816 +0000 UTC m=+146.209462651" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.484365 4761 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gh7tn container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.484400 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" podUID="427ade36-c2c8-4e77-8139-59f950c4368a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.512237 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" podStartSLOduration=125.512224244 podStartE2EDuration="2m5.512224244s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.511349421 +0000 UTC m=+146.244695266" watchObservedRunningTime="2025-11-25 22:59:10.512224244 +0000 UTC m=+146.245570079" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.517084 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" event={"ID":"d5da9ea9-b02e-4112-9f0c-910aa1bc2229","Type":"ContainerStarted","Data":"cd2cacefc1601c09ae0cdd26a2fe2e0bbc26cc8a2ca419bf1a90ca37a5d3254f"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.542894 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wlhcm" event={"ID":"019554bd-ec80-42eb-8439-fe453d8537cc","Type":"ContainerStarted","Data":"c000b515813c50a2f7b5d5c343af741a0800072546d62d40ac6dff475487b57b"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.557377 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.557977 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.057963132 +0000 UTC m=+146.791308967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.585285 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5jsv6" podStartSLOduration=125.585265979 podStartE2EDuration="2m5.585265979s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.556844405 +0000 UTC m=+146.290190250" watchObservedRunningTime="2025-11-25 22:59:10.585265979 +0000 UTC m=+146.318611814" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.585960 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" podStartSLOduration=125.585957216 podStartE2EDuration="2m5.585957216s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.584838608 +0000 UTC m=+146.318184443" watchObservedRunningTime="2025-11-25 22:59:10.585957216 +0000 UTC m=+146.319303051" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.591881 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" event={"ID":"2a3b9d72-ca02-45b2-ad09-63ad014e0eed","Type":"ContainerStarted","Data":"785c8e5a1f111b008cd2ba99ecd235789114017915a975838e815c89312edda1"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.591927 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" event={"ID":"2a3b9d72-ca02-45b2-ad09-63ad014e0eed","Type":"ContainerStarted","Data":"76746b9f7f945ddb894cbfa887eda0dd863f79412df39d4e94ea67fa437384ac"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.625927 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" event={"ID":"2a2a3111-ccac-4848-ae78-138daab3a2c6","Type":"ContainerStarted","Data":"3b3201ba47d13fcf77c3395e1c508be5790b2afb9ec3b848ed657af9e0c17071"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.642955 4761 generic.go:334] "Generic (PLEG): container finished" podID="7219fd35-3830-495d-ac04-9b65ad696259" containerID="142b78121c12fb967064b27181580af9a140ef6d368e15105e31a3b5d5980fda" exitCode=0 Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.643039 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" event={"ID":"7219fd35-3830-495d-ac04-9b65ad696259","Type":"ContainerDied","Data":"142b78121c12fb967064b27181580af9a140ef6d368e15105e31a3b5d5980fda"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.652132 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" event={"ID":"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7","Type":"ContainerStarted","Data":"dff4985b5f5837eba317385aa1e6b721099ed18cf307f3b2c57280c1905bc813"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.661015 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" event={"ID":"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58","Type":"ContainerStarted","Data":"e5ee7eba34bdec6c096bc9b811b4ef757bfddaeee3b6c1edaacb35513cb0a5e9"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.661222 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.661511 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.161499434 +0000 UTC m=+146.894845269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.667548 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4zv4" event={"ID":"33bd1f91-64e1-40d9-97e3-08ba2c52ea97","Type":"ContainerStarted","Data":"6ce8c7181d1baf6ff816946f49840d939ad0cddfb1c38bebbd074eb1057aa300"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.674944 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" event={"ID":"fd8e84c9-b40f-4e03-a91a-508fd653ee07","Type":"ContainerStarted","Data":"0368cb1288751616429b9f1059698c8f8e8c20b9e5a89cfda0fbeb12821bc4cb"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.676551 4761 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cqrtc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.676584 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.690524 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" podStartSLOduration=125.690509143 podStartE2EDuration="2m5.690509143s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.69039 +0000 UTC m=+146.423735845" watchObservedRunningTime="2025-11-25 22:59:10.690509143 +0000 UTC m=+146.423854978" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.690678 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-wlhcm" podStartSLOduration=125.690674117 podStartE2EDuration="2m5.690674117s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.614801761 +0000 UTC m=+146.348147616" watchObservedRunningTime="2025-11-25 22:59:10.690674117 +0000 UTC m=+146.424019952" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.694877 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v8ftk" event={"ID":"2244cddd-5bc9-46a0-86fe-64fa10c99ef1","Type":"ContainerStarted","Data":"cdb308a1102386c2dc8510992ba27143956d5052b63cd94eca45992f623560ce"} Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.748089 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5n2zp" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.748921 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" podStartSLOduration=125.74890158 podStartE2EDuration="2m5.74890158s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.727429741 +0000 UTC m=+146.460775596" watchObservedRunningTime="2025-11-25 22:59:10.74890158 +0000 UTC m=+146.482247405" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.750358 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" podStartSLOduration=124.750351837 podStartE2EDuration="2m4.750351837s" podCreationTimestamp="2025-11-25 22:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:10.748339076 +0000 UTC m=+146.481684921" watchObservedRunningTime="2025-11-25 22:59:10.750351837 +0000 UTC m=+146.483697692" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.768108 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.770387 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.270371969 +0000 UTC m=+147.003717804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.812732 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.813933 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.840171 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.869736 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.871319 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.371307855 +0000 UTC m=+147.104653690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.940918 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.946631 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:10 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:10 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:10 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.946715 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:10 crc kubenswrapper[4761]: I1125 22:59:10.971789 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:10 crc kubenswrapper[4761]: E1125 22:59:10.972095 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.472080717 +0000 UTC m=+147.205426552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.073380 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.073783 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.573765812 +0000 UTC m=+147.307111647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.174863 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.175121 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.675092818 +0000 UTC m=+147.408438653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.276690 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.277107 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.777091291 +0000 UTC m=+147.510437126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.377994 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.378181 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.8781598 +0000 UTC m=+147.611505635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.378689 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.379029 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.879021342 +0000 UTC m=+147.612367177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.480096 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.480285 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.980259015 +0000 UTC m=+147.713604850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.480365 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.480612 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:11.980605234 +0000 UTC m=+147.713951069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.581638 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.581827 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.081803467 +0000 UTC m=+147.815149302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.581916 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.582188 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.082175736 +0000 UTC m=+147.815521571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.682642 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.682814 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.182790604 +0000 UTC m=+147.916136439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.682905 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.683227 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.183211235 +0000 UTC m=+147.916557160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.698206 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" event={"ID":"06ae2ec0-c649-45b6-b762-6827b898a96b","Type":"ContainerStarted","Data":"96df3c50139b0c92dabbd6052ca18fc7171a929b058d5452797f90c846f12e65"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.699608 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-t87pc" event={"ID":"5f54a75a-fa05-4a82-b40f-eba6b9cc3af7","Type":"ContainerStarted","Data":"d6157c74fc51ce32b5a9c6c314e2f3a829aec08a1459d43d6323a6aa6b260041"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.701066 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" event={"ID":"4ed043b5-92aa-460d-8ced-1ac775019c08","Type":"ContainerStarted","Data":"5ac1c86ce599c124df3d2d59dba9e26eab990e259d9aeacb4c8199bb338d5f60"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.703066 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" event={"ID":"592cb417-405d-4a24-bd0b-b16f550e2f9f","Type":"ContainerStarted","Data":"1e9c905d6d5ec0155996a329faf1d4fdd959a131927853cecb80b8be46d18c44"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.705094 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" event={"ID":"7219fd35-3830-495d-ac04-9b65ad696259","Type":"ContainerStarted","Data":"df9c6e497fcbf397dc61ee8eea05768148e92344207937c0a4a89ab19b8fa6d8"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.705130 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" event={"ID":"7219fd35-3830-495d-ac04-9b65ad696259","Type":"ContainerStarted","Data":"ebe038c05d226187015e5c8d0fc6fed984cc3f7ecb292e060f783be3e345286d"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.706729 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" event={"ID":"18dc5416-2a42-4663-87dc-68c866e00af2","Type":"ContainerStarted","Data":"36f08d9eb55ba6cc42f81cf7e003caf7ff30bfeb76097be8e3803f062369fe16"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.706781 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.707742 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" event={"ID":"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3","Type":"ContainerStarted","Data":"e226dedb2b473e1e6997ad3f1d676f4caba84e619dc5d4da7886acd0e7d51a13"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.708615 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9xzkl" event={"ID":"d5da9ea9-b02e-4112-9f0c-910aa1bc2229","Type":"ContainerStarted","Data":"a85a935734a48adb7ef7d5e620378c5e98cf1c77b61e316f8069298f0076d98a"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.709519 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" event={"ID":"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58","Type":"ContainerStarted","Data":"73a56f32a8c417eb78115c31f6ff03cae0bb3357026cd7a8d53fb24b3f074798"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.710908 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" event={"ID":"1477d4f4-8b05-4eb9-a972-bcc2408a3c00","Type":"ContainerStarted","Data":"e1ec162ec2dbf7ed93d457ee4be8a42dd5b0ff443de82fb0deb043a67ebb1736"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.710937 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" event={"ID":"1477d4f4-8b05-4eb9-a972-bcc2408a3c00","Type":"ContainerStarted","Data":"4ed100552edb9cb03c3534920f71c1e085438e90ca771dd76ca2c1ad3b688211"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.712460 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" event={"ID":"55868319-fab2-4d45-8176-099029f717bb","Type":"ContainerStarted","Data":"858c8622ebaa909079fc0b981b41e821ddf64c327851c0541433c2d7759faf62"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.714266 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxvl7" event={"ID":"311fcb57-e225-42e2-8ba8-f21e7b83eb20","Type":"ContainerStarted","Data":"c945e177b515bde4f4c15e32dc6b1f8cfcd882b20f6580377bd58c00623827bc"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.714308 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxvl7" event={"ID":"311fcb57-e225-42e2-8ba8-f21e7b83eb20","Type":"ContainerStarted","Data":"b4dbecbd4da6a1de91756a8ee9c54e47c17447c4e03464de39d4d7bbe037589e"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.714393 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.716051 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" event={"ID":"2a3b9d72-ca02-45b2-ad09-63ad014e0eed","Type":"ContainerStarted","Data":"f4a5fe90c12aff156e60386be4fe0234289f02d92cf62673d677762802a740e1"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.717285 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-g6rkl" event={"ID":"2a2a3111-ccac-4848-ae78-138daab3a2c6","Type":"ContainerStarted","Data":"9b129c60f596a090aae553a4d8943c5c7e92d504e67597d6946e27468a09f244"} Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.718376 4761 patch_prober.go:28] interesting pod/downloads-7954f5f757-5jsv6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.718409 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5jsv6" podUID="f7fec219-569e-45e0-8cff-7ca1c78139ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.723969 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xzlxb" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.724260 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.724995 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d7lr4" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.733503 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.760337 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gh7tn" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.763906 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tm84l" podStartSLOduration=126.763888942 podStartE2EDuration="2m6.763888942s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:11.761371848 +0000 UTC m=+147.494717693" watchObservedRunningTime="2025-11-25 22:59:11.763888942 +0000 UTC m=+147.497234787" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.784054 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.784227 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.284197282 +0000 UTC m=+148.017543117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.785327 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.787647 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.287638079 +0000 UTC m=+148.020983914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.807541 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wqpjh" podStartSLOduration=126.807526948 podStartE2EDuration="2m6.807526948s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:11.806850831 +0000 UTC m=+147.540196676" watchObservedRunningTime="2025-11-25 22:59:11.807526948 +0000 UTC m=+147.540872783" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.883478 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-j4dlq" podStartSLOduration=126.883459036 podStartE2EDuration="2m6.883459036s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:11.883311532 +0000 UTC m=+147.616657387" watchObservedRunningTime="2025-11-25 22:59:11.883459036 +0000 UTC m=+147.616804871" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.886455 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.888139 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.388117943 +0000 UTC m=+148.121463778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.923368 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pr4rh" podStartSLOduration=126.92334983800001 podStartE2EDuration="2m6.923349838s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:11.921967424 +0000 UTC m=+147.655313259" watchObservedRunningTime="2025-11-25 22:59:11.923349838 +0000 UTC m=+147.656695663" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.939597 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:11 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:11 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:11 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.939651 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.958488 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" podStartSLOduration=125.958468461 podStartE2EDuration="2m5.958468461s" podCreationTimestamp="2025-11-25 22:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:11.952324546 +0000 UTC m=+147.685670381" watchObservedRunningTime="2025-11-25 22:59:11.958468461 +0000 UTC m=+147.691814296" Nov 25 22:59:11 crc kubenswrapper[4761]: I1125 22:59:11.987827 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:11 crc kubenswrapper[4761]: E1125 22:59:11.988147 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.488135646 +0000 UTC m=+148.221481481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.089386 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.089575 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.589547934 +0000 UTC m=+148.322893769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.089799 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.090108 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.590100348 +0000 UTC m=+148.323446183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.095177 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-66j2w" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.135228 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.135277 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.143980 4761 patch_prober.go:28] interesting pod/apiserver-76f77b778f-n5snq container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.144063 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" podUID="7219fd35-3830-495d-ac04-9b65ad696259" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.190181 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.190559 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.690545922 +0000 UTC m=+148.423891757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.193133 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nwpc5" podStartSLOduration=127.193121696 podStartE2EDuration="2m7.193121696s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:12.187668959 +0000 UTC m=+147.921014804" watchObservedRunningTime="2025-11-25 22:59:12.193121696 +0000 UTC m=+147.926467531" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.297402 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.297986 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.79797465 +0000 UTC m=+148.531320485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.340570 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" podStartSLOduration=127.340549581 podStartE2EDuration="2m7.340549581s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:12.339968036 +0000 UTC m=+148.073313871" watchObservedRunningTime="2025-11-25 22:59:12.340549581 +0000 UTC m=+148.073895416" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.341225 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" podStartSLOduration=127.341220617 podStartE2EDuration="2m7.341220617s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:12.325779069 +0000 UTC m=+148.059124904" watchObservedRunningTime="2025-11-25 22:59:12.341220617 +0000 UTC m=+148.074566452" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.398288 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.401459 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vxvl7" podStartSLOduration=8.40144704 podStartE2EDuration="8.40144704s" podCreationTimestamp="2025-11-25 22:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:12.371465187 +0000 UTC m=+148.104811032" watchObservedRunningTime="2025-11-25 22:59:12.40144704 +0000 UTC m=+148.134792875" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.402175 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:12.902162809 +0000 UTC m=+148.635508644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.440180 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.441128 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.457507 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.465127 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.502511 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.502630 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfk8p\" (UniqueName: \"kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.502687 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.502822 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.503289 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.003268229 +0000 UTC m=+148.736614074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.603970 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.604165 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.104133333 +0000 UTC m=+148.837479168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.604376 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.604418 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfk8p\" (UniqueName: \"kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.604887 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.604936 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.605020 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.605179 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.605342 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.105334733 +0000 UTC m=+148.838680568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.610207 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.611156 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.615325 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.626739 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfk8p\" (UniqueName: \"kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p\") pod \"certified-operators-rdcmf\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.633773 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.673348 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.706127 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.706331 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.706380 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.706421 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h2l9\" (UniqueName: \"kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.706546 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.206530836 +0000 UTC m=+148.939876671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.724047 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" event={"ID":"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58","Type":"ContainerStarted","Data":"419db7cbf5af0ca2ef2228a6e5f2d5cc16648f55dddd5fcec4f2f0d16562fa52"} Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.724747 4761 patch_prober.go:28] interesting pod/downloads-7954f5f757-5jsv6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.724776 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5jsv6" podUID="f7fec219-569e-45e0-8cff-7ca1c78139ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.798838 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.809809 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.810679 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.817153 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.817463 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.817635 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h2l9\" (UniqueName: \"kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.817726 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.821821 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.829892 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.329877425 +0000 UTC m=+149.063223260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.830730 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.846395 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.885460 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h2l9\" (UniqueName: \"kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9\") pod \"community-operators-gss6x\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.923122 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.923350 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.923448 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98ltz\" (UniqueName: \"kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.923481 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:12 crc kubenswrapper[4761]: E1125 22:59:12.923581 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.423566509 +0000 UTC m=+149.156912344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.923786 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gss6x" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.943843 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:12 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:12 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:12 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.943890 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:12 crc kubenswrapper[4761]: I1125 22:59:12.954184 4761 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.006811 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.007929 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.013489 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.025021 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.025065 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98ltz\" (UniqueName: \"kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.025094 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.025127 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.025945 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: E1125 22:59:13.026172 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.526161646 +0000 UTC m=+149.259507481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t6smk" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.026690 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.059486 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98ltz\" (UniqueName: \"kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz\") pod \"certified-operators-kfwx7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.128198 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.128369 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.128442 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l4l8\" (UniqueName: \"kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.128474 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: E1125 22:59:13.128563 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 22:59:13.628547308 +0000 UTC m=+149.361893143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.170781 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.171182 4761 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T22:59:12.954211349Z","Handler":null,"Name":""} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.174541 4761 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.174566 4761 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.230434 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l4l8\" (UniqueName: \"kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.230483 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.230529 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.230563 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.231305 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.231503 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.249326 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l4l8\" (UniqueName: \"kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8\") pod \"community-operators-9jt5l\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.265369 4761 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.265403 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.279294 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.351044 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.375080 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t6smk\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.425234 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fkj42" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.434229 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.486502 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.584722 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.678810 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.728691 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.768744 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerStarted","Data":"70ba23c2476039f095a3627256155af6213feea96e2812b5eefb8d002119d036"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.797474 4761 generic.go:334] "Generic (PLEG): container finished" podID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerID="10b195a7b95dddb251cec4757fc93f79f88c56cc4ac0ceea4cce735772949a20" exitCode=0 Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.797560 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerDied","Data":"10b195a7b95dddb251cec4757fc93f79f88c56cc4ac0ceea4cce735772949a20"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.797585 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerStarted","Data":"dbdf8ca9418dbeebeeeea1b17cc626d27f80cd27b7fc7920abaf409c24523d9f"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.799879 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.802930 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerStarted","Data":"f33ab398643901f3bb54786db32d71e65fb672ba8c9d4471999762229045ec7f"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.848023 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" event={"ID":"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58","Type":"ContainerStarted","Data":"973ec3c4e413da21381747a3b0bbfabadb49efde42c3ec64da0ea9d70d474b34"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.848347 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.849273 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" event={"ID":"9013dc7b-48e2-4c1b-9ca3-4ed3af15bd58","Type":"ContainerStarted","Data":"f3826bfbe33ddfbdd078ee38a1cef2ed832cce84c401e9714933cfbcf250c99d"} Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.849318 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.851367 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.851788 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.860311 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.915626 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jrd27" podStartSLOduration=9.915607424 podStartE2EDuration="9.915607424s" podCreationTimestamp="2025-11-25 22:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:13.913201003 +0000 UTC m=+149.646546838" watchObservedRunningTime="2025-11-25 22:59:13.915607424 +0000 UTC m=+149.648953259" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.941656 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.941811 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.941847 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.941873 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.941900 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.942078 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.945644 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:13 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:13 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:13 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.945708 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.947797 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.956305 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.959583 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:59:13 crc kubenswrapper[4761]: I1125 22:59:13.970986 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.043263 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.043310 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.043640 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.052948 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.063141 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.063940 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.070078 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.114677 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 22:59:14 crc kubenswrapper[4761]: W1125 22:59:14.137968 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc80d2ace_7b69_4960_a664_41fb94d072de.slice/crio-c9824be5f6dfd9a447c2c47a0ae058527af245c0b42ff13d8615c71f7589782f WatchSource:0}: Error finding container c9824be5f6dfd9a447c2c47a0ae058527af245c0b42ff13d8615c71f7589782f: Status 404 returned error can't find the container with id c9824be5f6dfd9a447c2c47a0ae058527af245c0b42ff13d8615c71f7589782f Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.182557 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.233531 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 22:59:14 crc kubenswrapper[4761]: W1125 22:59:14.730347 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e68814c1eefc34cd932ac599049fd0397c793bd4a1050540e174cd8ff4c960ac WatchSource:0}: Error finding container e68814c1eefc34cd932ac599049fd0397c793bd4a1050540e174cd8ff4c960ac: Status 404 returned error can't find the container with id e68814c1eefc34cd932ac599049fd0397c793bd4a1050540e174cd8ff4c960ac Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.734690 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.793899 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.794836 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.796539 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.807860 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.857136 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.857750 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.857798 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvsx\" (UniqueName: \"kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: W1125 22:59:14.901970 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b37d4b355e7245b90266fd9de0e22abf884e75f5d6f44f2ee147c5940a3e2616 WatchSource:0}: Error finding container b37d4b355e7245b90266fd9de0e22abf884e75f5d6f44f2ee147c5940a3e2616: Status 404 returned error can't find the container with id b37d4b355e7245b90266fd9de0e22abf884e75f5d6f44f2ee147c5940a3e2616 Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.903440 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" event={"ID":"c80d2ace-7b69-4960-a664-41fb94d072de","Type":"ContainerStarted","Data":"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.903492 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" event={"ID":"c80d2ace-7b69-4960-a664-41fb94d072de","Type":"ContainerStarted","Data":"c9824be5f6dfd9a447c2c47a0ae058527af245c0b42ff13d8615c71f7589782f"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.903635 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.908760 4761 generic.go:334] "Generic (PLEG): container finished" podID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerID="1c350e067b76d01616548042ffe1b22dd7e6322fd5dff8554887ccdaadffa573" exitCode=0 Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.908807 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerDied","Data":"1c350e067b76d01616548042ffe1b22dd7e6322fd5dff8554887ccdaadffa573"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.913100 4761 generic.go:334] "Generic (PLEG): container finished" podID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerID="30af171f5ea6ccf5e281bf19a85b0e3d31109da1ab7699711865102684d0b6e6" exitCode=0 Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.913161 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerDied","Data":"30af171f5ea6ccf5e281bf19a85b0e3d31109da1ab7699711865102684d0b6e6"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.915551 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e","Type":"ContainerStarted","Data":"f303ef403cb6601e32bcbe160e81a5bbf2be7a0ec4c6bdad5a7c76464dfbe90e"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.922238 4761 generic.go:334] "Generic (PLEG): container finished" podID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerID="d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c" exitCode=0 Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.925734 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerDied","Data":"d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.925787 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerStarted","Data":"78908cdd691aa51cf4fe0f8639cbaad8cb3a35254afbd960e5d77a062d88db8c"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.930169 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"017e9b6a6d4fc2c72628abd8940c624f4f748aab27c2d4bbd167fe92e56ba43d"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.935608 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e68814c1eefc34cd932ac599049fd0397c793bd4a1050540e174cd8ff4c960ac"} Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.939766 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:14 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:14 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:14 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.939883 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.946496 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" podStartSLOduration=129.946477175 podStartE2EDuration="2m9.946477175s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:14.925622741 +0000 UTC m=+150.658968576" watchObservedRunningTime="2025-11-25 22:59:14.946477175 +0000 UTC m=+150.679823010" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.958709 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.958792 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.958817 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvsx\" (UniqueName: \"kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.959359 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.960424 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:14 crc kubenswrapper[4761]: I1125 22:59:14.980841 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvsx\" (UniqueName: \"kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx\") pod \"redhat-marketplace-56j56\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.021145 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.114867 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.197955 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.199181 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.230412 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.263275 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.263353 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28fjn\" (UniqueName: \"kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.263386 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.352804 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.365607 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.365676 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28fjn\" (UniqueName: \"kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.365734 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.366148 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.366160 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: W1125 22:59:15.369504 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod363875a6_dc39_4220_ab62_d8f390aa0625.slice/crio-c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a WatchSource:0}: Error finding container c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a: Status 404 returned error can't find the container with id c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.397231 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28fjn\" (UniqueName: \"kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn\") pod \"redhat-marketplace-87tks\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.546913 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.597875 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.599219 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.605735 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.607308 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.669513 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmf42\" (UniqueName: \"kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.669567 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.669596 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.770810 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmf42\" (UniqueName: \"kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.770861 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.770958 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.772764 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.773143 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.790553 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.799463 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.800429 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.809199 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.828300 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmf42\" (UniqueName: \"kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42\") pod \"redhat-operators-4crxm\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.872276 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.872378 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.872410 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r99hn\" (UniqueName: \"kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.939325 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:15 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:15 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:15 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.939417 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.939544 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.948177 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0f504a6b55e4efd2d45ff8edfa66ef98b9ef5f2d431e3be93d8642c904d46e1c"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.948267 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.955016 4761 generic.go:334] "Generic (PLEG): container finished" podID="363875a6-dc39-4220-ab62-d8f390aa0625" containerID="629acc97814673915258cd5a4f750e64e214bda7dcd3e94a543f9a3c0fbe824a" exitCode=0 Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.955097 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerDied","Data":"629acc97814673915258cd5a4f750e64e214bda7dcd3e94a543f9a3c0fbe824a"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.955131 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerStarted","Data":"c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.966638 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"77450c12eb8c1341cf75be1dad826e6afaf85e38200ea3edb33e7d011509e15e"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.966686 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b37d4b355e7245b90266fd9de0e22abf884e75f5d6f44f2ee147c5940a3e2616"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.973117 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerStarted","Data":"9196419038ba51312cc6d7f937bb86db7a1e02cb7bd420b8f6430aadb15b5d38"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.973963 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.974068 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.974099 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r99hn\" (UniqueName: \"kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.974859 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.976674 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.984488 4761 generic.go:334] "Generic (PLEG): container finished" podID="3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" containerID="ba714140179a845e7053b42c8345173067679109f69382356a025470e542a805" exitCode=0 Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.984629 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e","Type":"ContainerDied","Data":"ba714140179a845e7053b42c8345173067679109f69382356a025470e542a805"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.989434 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"056e137aed57b5d464e9163bfbd5e76e04e9c440bd4c5309e4e59bcca0af1c4c"} Nov 25 22:59:15 crc kubenswrapper[4761]: I1125 22:59:15.995782 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r99hn\" (UniqueName: \"kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn\") pod \"redhat-operators-j9mf6\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:16 crc kubenswrapper[4761]: I1125 22:59:16.190565 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 22:59:16 crc kubenswrapper[4761]: I1125 22:59:16.283620 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 22:59:16 crc kubenswrapper[4761]: W1125 22:59:16.329477 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a915718_fdf1_4829_9d76_a704fe9cd833.slice/crio-866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a WatchSource:0}: Error finding container 866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a: Status 404 returned error can't find the container with id 866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a Nov 25 22:59:16 crc kubenswrapper[4761]: I1125 22:59:16.494247 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 22:59:16 crc kubenswrapper[4761]: I1125 22:59:16.939370 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:16 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:16 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:16 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:16 crc kubenswrapper[4761]: I1125 22:59:16.939423 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.016013 4761 generic.go:334] "Generic (PLEG): container finished" podID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerID="21623d16205bbd6355c0661eb8dd3c24d0f523f6eb827640119a1907ff7eddfb" exitCode=0 Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.020815 4761 generic.go:334] "Generic (PLEG): container finished" podID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerID="1b14595e2051e6e9317c43cd6ff8719f252a761f9e467e31376252c544d639c0" exitCode=0 Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.025436 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerDied","Data":"21623d16205bbd6355c0661eb8dd3c24d0f523f6eb827640119a1907ff7eddfb"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.025606 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerStarted","Data":"c42631337247f8759f761dd3bbe28da23dceaa6645f75dcd77d4e2806fd42974"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.025617 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerDied","Data":"1b14595e2051e6e9317c43cd6ff8719f252a761f9e467e31376252c544d639c0"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.025627 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerStarted","Data":"866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.037625 4761 generic.go:334] "Generic (PLEG): container finished" podID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerID="70ed616a2f61b8c050f1064a8506846cacfeb9a3ca2ddb7fc8ce0dc8d9411f1f" exitCode=0 Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.037734 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerDied","Data":"70ed616a2f61b8c050f1064a8506846cacfeb9a3ca2ddb7fc8ce0dc8d9411f1f"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.045301 4761 generic.go:334] "Generic (PLEG): container finished" podID="6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" containerID="e226dedb2b473e1e6997ad3f1d676f4caba84e619dc5d4da7886acd0e7d51a13" exitCode=0 Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.045412 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" event={"ID":"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3","Type":"ContainerDied","Data":"e226dedb2b473e1e6997ad3f1d676f4caba84e619dc5d4da7886acd0e7d51a13"} Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.141122 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.147848 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-n5snq" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.315771 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.480173 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.480446 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.482384 4761 patch_prober.go:28] interesting pod/console-f9d7485db-g4wpz container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.482432 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-g4wpz" podUID="b72410eb-f8eb-40c5-ae89-ba009453384c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.509007 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir\") pod \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.509061 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access\") pod \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\" (UID: \"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e\") " Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.510591 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" (UID: "3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.597690 4761 patch_prober.go:28] interesting pod/downloads-7954f5f757-5jsv6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.597735 4761 patch_prober.go:28] interesting pod/downloads-7954f5f757-5jsv6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.597760 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5jsv6" podUID="f7fec219-569e-45e0-8cff-7ca1c78139ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.597760 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5jsv6" podUID="f7fec219-569e-45e0-8cff-7ca1c78139ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.611437 4761 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.889736 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" (UID: "3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.916009 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.936605 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.938541 4761 patch_prober.go:28] interesting pod/router-default-5444994796-wlhcm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 22:59:17 crc kubenswrapper[4761]: [-]has-synced failed: reason withheld Nov 25 22:59:17 crc kubenswrapper[4761]: [+]process-running ok Nov 25 22:59:17 crc kubenswrapper[4761]: healthz check failed Nov 25 22:59:17 crc kubenswrapper[4761]: I1125 22:59:17.938598 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wlhcm" podUID="019554bd-ec80-42eb-8439-fe453d8537cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.054402 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e","Type":"ContainerDied","Data":"f303ef403cb6601e32bcbe160e81a5bbf2be7a0ec4c6bdad5a7c76464dfbe90e"} Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.054441 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f303ef403cb6601e32bcbe160e81a5bbf2be7a0ec4c6bdad5a7c76464dfbe90e" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.055161 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.322973 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.431088 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r9jm\" (UniqueName: \"kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm\") pod \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.431146 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume\") pod \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.431183 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume\") pod \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\" (UID: \"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3\") " Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.432147 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume" (OuterVolumeSpecName: "config-volume") pod "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" (UID: "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.446291 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm" (OuterVolumeSpecName: "kube-api-access-2r9jm") pod "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" (UID: "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3"). InnerVolumeSpecName "kube-api-access-2r9jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.446441 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" (UID: "6c27c67d-99bf-4eea-bf3d-96a034d1e1b3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.534229 4761 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.534849 4761 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.534865 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r9jm\" (UniqueName: \"kubernetes.io/projected/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3-kube-api-access-2r9jm\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.938516 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:18 crc kubenswrapper[4761]: I1125 22:59:18.940753 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-wlhcm" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.065990 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" event={"ID":"6c27c67d-99bf-4eea-bf3d-96a034d1e1b3","Type":"ContainerDied","Data":"2d492fa03dd4556dc52304c02765ee0c5c3fb367f85b35fe2ac0dcf12a26ff9f"} Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.066062 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d492fa03dd4556dc52304c02765ee0c5c3fb367f85b35fe2ac0dcf12a26ff9f" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.066068 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.203173 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 22:59:19 crc kubenswrapper[4761]: E1125 22:59:19.203457 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" containerName="pruner" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.203472 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" containerName="pruner" Nov 25 22:59:19 crc kubenswrapper[4761]: E1125 22:59:19.203484 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" containerName="collect-profiles" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.203490 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" containerName="collect-profiles" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.203591 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="3baa6d2d-d7ac-4e4b-b84d-9ca2935f8c7e" containerName="pruner" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.203633 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" containerName="collect-profiles" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.204217 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.210104 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.210251 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.210418 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.242415 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.242521 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.343316 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.343389 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.343676 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.359041 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:19 crc kubenswrapper[4761]: I1125 22:59:19.547165 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:21 crc kubenswrapper[4761]: I1125 22:59:21.128232 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:59:21 crc kubenswrapper[4761]: I1125 22:59:21.128577 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:59:22 crc kubenswrapper[4761]: I1125 22:59:22.764202 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vxvl7" Nov 25 22:59:27 crc kubenswrapper[4761]: I1125 22:59:27.487446 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:27 crc kubenswrapper[4761]: I1125 22:59:27.493485 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-g4wpz" Nov 25 22:59:27 crc kubenswrapper[4761]: I1125 22:59:27.612006 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5jsv6" Nov 25 22:59:28 crc kubenswrapper[4761]: I1125 22:59:28.994867 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:59:29 crc kubenswrapper[4761]: I1125 22:59:29.004560 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d99b4cb-ae29-4198-a407-a1cfe211f7a3-metrics-certs\") pod \"network-metrics-daemon-lhpwb\" (UID: \"2d99b4cb-ae29-4198-a407-a1cfe211f7a3\") " pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:59:29 crc kubenswrapper[4761]: I1125 22:59:29.036223 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhpwb" Nov 25 22:59:29 crc kubenswrapper[4761]: I1125 22:59:29.493566 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 22:59:29 crc kubenswrapper[4761]: I1125 22:59:29.546503 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lhpwb"] Nov 25 22:59:33 crc kubenswrapper[4761]: I1125 22:59:33.494299 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 22:59:45 crc kubenswrapper[4761]: W1125 22:59:45.991959 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0a6ba148_70a2_4be4_9f84_1bdf1f121d95.slice/crio-c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510 WatchSource:0}: Error finding container c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510: Status 404 returned error can't find the container with id c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510 Nov 25 22:59:45 crc kubenswrapper[4761]: W1125 22:59:45.993458 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d99b4cb_ae29_4198_a407_a1cfe211f7a3.slice/crio-b5b92b813f0b95bc8202732d50e2f7814026a3c4dce4f08c1bd57bf5f1e6243b WatchSource:0}: Error finding container b5b92b813f0b95bc8202732d50e2f7814026a3c4dce4f08c1bd57bf5f1e6243b: Status 404 returned error can't find the container with id b5b92b813f0b95bc8202732d50e2f7814026a3c4dce4f08c1bd57bf5f1e6243b Nov 25 22:59:46 crc kubenswrapper[4761]: I1125 22:59:46.245086 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0a6ba148-70a2-4be4-9f84-1bdf1f121d95","Type":"ContainerStarted","Data":"c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510"} Nov 25 22:59:46 crc kubenswrapper[4761]: I1125 22:59:46.266366 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" event={"ID":"2d99b4cb-ae29-4198-a407-a1cfe211f7a3","Type":"ContainerStarted","Data":"b5b92b813f0b95bc8202732d50e2f7814026a3c4dce4f08c1bd57bf5f1e6243b"} Nov 25 22:59:47 crc kubenswrapper[4761]: I1125 22:59:47.931418 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5rncl" Nov 25 22:59:51 crc kubenswrapper[4761]: I1125 22:59:51.128373 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:59:51 crc kubenswrapper[4761]: I1125 22:59:51.128920 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:59:51 crc kubenswrapper[4761]: E1125 22:59:51.747594 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 22:59:51 crc kubenswrapper[4761]: E1125 22:59:51.747885 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmf42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4crxm_openshift-marketplace(7a915718-fdf1-4829-9d76-a704fe9cd833): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:51 crc kubenswrapper[4761]: E1125 22:59:51.749265 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4crxm" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.921811 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.922280 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dfk8p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rdcmf_openshift-marketplace(33e7b5be-2cc4-4598-b197-e34a5c3d4e0a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.923682 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rdcmf" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.951296 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.951538 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r99hn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-j9mf6_openshift-marketplace(9b235fa2-d6a8-41b9-ab55-41fd4a60c90d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:52 crc kubenswrapper[4761]: E1125 22:59:52.953569 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-j9mf6" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" Nov 25 22:59:54 crc kubenswrapper[4761]: I1125 22:59:54.067474 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.450596 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4crxm" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.450852 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rdcmf" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.450928 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-j9mf6" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.515135 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.515288 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9h2l9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gss6x_openshift-marketplace(a310e19b-3b47-4c9c-ba97-9e4761d75fa9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.516766 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gss6x" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.537925 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.538088 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5l4l8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9jt5l_openshift-marketplace(156832a0-c911-4ccf-bb41-511cbcdaab8c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:54 crc kubenswrapper[4761]: E1125 22:59:54.539489 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9jt5l" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.418616 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gss6x" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.418864 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9jt5l" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.479338 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.479522 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ggvsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-56j56_openshift-marketplace(363875a6-dc39-4220-ab62-d8f390aa0625): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.481001 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-56j56" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.516165 4761 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.516329 4761 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-28fjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-87tks_openshift-marketplace(8943ee05-ad49-4fe8-b4e2-104aac3fd2c2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 22:59:55 crc kubenswrapper[4761]: E1125 22:59:55.517482 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-87tks" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.371142 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0a6ba148-70a2-4be4-9f84-1bdf1f121d95","Type":"ContainerStarted","Data":"87e95cebc303e1bef201efd673dad3cb7392286835beab0455ca23c466fbf013"} Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.374080 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" event={"ID":"2d99b4cb-ae29-4198-a407-a1cfe211f7a3","Type":"ContainerStarted","Data":"d758c6451a9877214ce6bedd578318ab8ed3683554c8b062e3aeeed8d2aeb621"} Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.374130 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhpwb" event={"ID":"2d99b4cb-ae29-4198-a407-a1cfe211f7a3","Type":"ContainerStarted","Data":"10121a6b135822d3506f32a6ba3ef3e9855d610d24d6902e24584d54b6e956e6"} Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.381164 4761 generic.go:334] "Generic (PLEG): container finished" podID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerID="0ead4f7bd5c80329344da831676280dfadd90f195d2e96dbe55ace1b64e71d03" exitCode=0 Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.381375 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerDied","Data":"0ead4f7bd5c80329344da831676280dfadd90f195d2e96dbe55ace1b64e71d03"} Nov 25 22:59:56 crc kubenswrapper[4761]: E1125 22:59:56.384922 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-87tks" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" Nov 25 22:59:56 crc kubenswrapper[4761]: E1125 22:59:56.385618 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-56j56" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.405511 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=37.405491449 podStartE2EDuration="37.405491449s" podCreationTimestamp="2025-11-25 22:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:56.402899694 +0000 UTC m=+192.136245539" watchObservedRunningTime="2025-11-25 22:59:56.405491449 +0000 UTC m=+192.138837294" Nov 25 22:59:56 crc kubenswrapper[4761]: I1125 22:59:56.456026 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lhpwb" podStartSLOduration=171.456003258 podStartE2EDuration="2m51.456003258s" podCreationTimestamp="2025-11-25 22:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:59:56.454599313 +0000 UTC m=+192.187945158" watchObservedRunningTime="2025-11-25 22:59:56.456003258 +0000 UTC m=+192.189349103" Nov 25 22:59:57 crc kubenswrapper[4761]: I1125 22:59:57.393732 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerStarted","Data":"064c2716a216e98a952041bfc5ac72306c37a8ddf4cf000246f6b3117ad3cfff"} Nov 25 22:59:57 crc kubenswrapper[4761]: I1125 22:59:57.397027 4761 generic.go:334] "Generic (PLEG): container finished" podID="0a6ba148-70a2-4be4-9f84-1bdf1f121d95" containerID="87e95cebc303e1bef201efd673dad3cb7392286835beab0455ca23c466fbf013" exitCode=0 Nov 25 22:59:57 crc kubenswrapper[4761]: I1125 22:59:57.397629 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0a6ba148-70a2-4be4-9f84-1bdf1f121d95","Type":"ContainerDied","Data":"87e95cebc303e1bef201efd673dad3cb7392286835beab0455ca23c466fbf013"} Nov 25 22:59:57 crc kubenswrapper[4761]: I1125 22:59:57.421524 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfwx7" podStartSLOduration=3.488785799 podStartE2EDuration="45.421500926s" podCreationTimestamp="2025-11-25 22:59:12 +0000 UTC" firstStartedPulling="2025-11-25 22:59:14.910791267 +0000 UTC m=+150.644137102" lastFinishedPulling="2025-11-25 22:59:56.843506354 +0000 UTC m=+192.576852229" observedRunningTime="2025-11-25 22:59:57.416171462 +0000 UTC m=+193.149517327" watchObservedRunningTime="2025-11-25 22:59:57.421500926 +0000 UTC m=+193.154846781" Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.684187 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.793910 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access\") pod \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.793984 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir\") pod \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\" (UID: \"0a6ba148-70a2-4be4-9f84-1bdf1f121d95\") " Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.794107 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0a6ba148-70a2-4be4-9f84-1bdf1f121d95" (UID: "0a6ba148-70a2-4be4-9f84-1bdf1f121d95"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.794311 4761 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.804936 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0a6ba148-70a2-4be4-9f84-1bdf1f121d95" (UID: "0a6ba148-70a2-4be4-9f84-1bdf1f121d95"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:59:58 crc kubenswrapper[4761]: I1125 22:59:58.895939 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a6ba148-70a2-4be4-9f84-1bdf1f121d95-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:59:59 crc kubenswrapper[4761]: I1125 22:59:59.409231 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0a6ba148-70a2-4be4-9f84-1bdf1f121d95","Type":"ContainerDied","Data":"c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510"} Nov 25 22:59:59 crc kubenswrapper[4761]: I1125 22:59:59.409680 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c09ec95025187eae41dc3d14fa744c49b5853485576864964cd3d8019e951510" Nov 25 22:59:59 crc kubenswrapper[4761]: I1125 22:59:59.409332 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.145043 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx"] Nov 25 23:00:00 crc kubenswrapper[4761]: E1125 23:00:00.145284 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a6ba148-70a2-4be4-9f84-1bdf1f121d95" containerName="pruner" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.145296 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a6ba148-70a2-4be4-9f84-1bdf1f121d95" containerName="pruner" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.145416 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a6ba148-70a2-4be4-9f84-1bdf1f121d95" containerName="pruner" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.145852 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.148087 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.148214 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.159541 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx"] Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.223691 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmcn\" (UniqueName: \"kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.223799 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.223862 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.325636 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmcn\" (UniqueName: \"kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.325763 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.325940 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.326656 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.343263 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.348902 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmcn\" (UniqueName: \"kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn\") pod \"collect-profiles-29401860-rxmjx\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.465989 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:00 crc kubenswrapper[4761]: I1125 23:00:00.681017 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx"] Nov 25 23:00:00 crc kubenswrapper[4761]: W1125 23:00:00.691433 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0984f952_54df_438b_929a_bb5ad66c2025.slice/crio-829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5 WatchSource:0}: Error finding container 829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5: Status 404 returned error can't find the container with id 829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5 Nov 25 23:00:01 crc kubenswrapper[4761]: I1125 23:00:01.419643 4761 generic.go:334] "Generic (PLEG): container finished" podID="0984f952-54df-438b-929a-bb5ad66c2025" containerID="29cdaab491a98caad217ba05340de1ea82f1726f580593316a1e1db6db127587" exitCode=0 Nov 25 23:00:01 crc kubenswrapper[4761]: I1125 23:00:01.419748 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" event={"ID":"0984f952-54df-438b-929a-bb5ad66c2025","Type":"ContainerDied","Data":"29cdaab491a98caad217ba05340de1ea82f1726f580593316a1e1db6db127587"} Nov 25 23:00:01 crc kubenswrapper[4761]: I1125 23:00:01.419983 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" event={"ID":"0984f952-54df-438b-929a-bb5ad66c2025","Type":"ContainerStarted","Data":"829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5"} Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.677460 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.763341 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume\") pod \"0984f952-54df-438b-929a-bb5ad66c2025\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.763419 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjmcn\" (UniqueName: \"kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn\") pod \"0984f952-54df-438b-929a-bb5ad66c2025\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.763495 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume\") pod \"0984f952-54df-438b-929a-bb5ad66c2025\" (UID: \"0984f952-54df-438b-929a-bb5ad66c2025\") " Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.764320 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume" (OuterVolumeSpecName: "config-volume") pod "0984f952-54df-438b-929a-bb5ad66c2025" (UID: "0984f952-54df-438b-929a-bb5ad66c2025"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.770559 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0984f952-54df-438b-929a-bb5ad66c2025" (UID: "0984f952-54df-438b-929a-bb5ad66c2025"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.776628 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn" (OuterVolumeSpecName: "kube-api-access-cjmcn") pod "0984f952-54df-438b-929a-bb5ad66c2025" (UID: "0984f952-54df-438b-929a-bb5ad66c2025"). InnerVolumeSpecName "kube-api-access-cjmcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.865371 4761 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0984f952-54df-438b-929a-bb5ad66c2025-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.865406 4761 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0984f952-54df-438b-929a-bb5ad66c2025-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:02 crc kubenswrapper[4761]: I1125 23:00:02.865419 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjmcn\" (UniqueName: \"kubernetes.io/projected/0984f952-54df-438b-929a-bb5ad66c2025-kube-api-access-cjmcn\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.171571 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.171633 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.332136 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.431036 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" event={"ID":"0984f952-54df-438b-929a-bb5ad66c2025","Type":"ContainerDied","Data":"829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5"} Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.431073 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401860-rxmjx" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.431110 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="829babd3bf83f646b82c59afa28074befc162cd6ee4235a42be37c2516ab49e5" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.485321 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:03 crc kubenswrapper[4761]: I1125 23:00:03.558779 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 23:00:05 crc kubenswrapper[4761]: I1125 23:00:05.444284 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfwx7" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="registry-server" containerID="cri-o://064c2716a216e98a952041bfc5ac72306c37a8ddf4cf000246f6b3117ad3cfff" gracePeriod=2 Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.452122 4761 generic.go:334] "Generic (PLEG): container finished" podID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerID="064c2716a216e98a952041bfc5ac72306c37a8ddf4cf000246f6b3117ad3cfff" exitCode=0 Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.452307 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerDied","Data":"064c2716a216e98a952041bfc5ac72306c37a8ddf4cf000246f6b3117ad3cfff"} Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.455061 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerStarted","Data":"d3ac805418d73f39bd407bc88b0ef28c7e96dc90cacb168af86acb2d86c25715"} Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.578277 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.619809 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities\") pod \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.619856 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98ltz\" (UniqueName: \"kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz\") pod \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.619884 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content\") pod \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\" (UID: \"000888e9-b922-46cb-a11a-baf8ec7fa8f7\") " Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.620723 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities" (OuterVolumeSpecName: "utilities") pod "000888e9-b922-46cb-a11a-baf8ec7fa8f7" (UID: "000888e9-b922-46cb-a11a-baf8ec7fa8f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.627383 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz" (OuterVolumeSpecName: "kube-api-access-98ltz") pod "000888e9-b922-46cb-a11a-baf8ec7fa8f7" (UID: "000888e9-b922-46cb-a11a-baf8ec7fa8f7"). InnerVolumeSpecName "kube-api-access-98ltz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.721675 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.721754 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98ltz\" (UniqueName: \"kubernetes.io/projected/000888e9-b922-46cb-a11a-baf8ec7fa8f7-kube-api-access-98ltz\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.908519 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "000888e9-b922-46cb-a11a-baf8ec7fa8f7" (UID: "000888e9-b922-46cb-a11a-baf8ec7fa8f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:06 crc kubenswrapper[4761]: I1125 23:00:06.923278 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000888e9-b922-46cb-a11a-baf8ec7fa8f7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.461991 4761 generic.go:334] "Generic (PLEG): container finished" podID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerID="d3ac805418d73f39bd407bc88b0ef28c7e96dc90cacb168af86acb2d86c25715" exitCode=0 Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.462069 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerDied","Data":"d3ac805418d73f39bd407bc88b0ef28c7e96dc90cacb168af86acb2d86c25715"} Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.466064 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwx7" event={"ID":"000888e9-b922-46cb-a11a-baf8ec7fa8f7","Type":"ContainerDied","Data":"70ba23c2476039f095a3627256155af6213feea96e2812b5eefb8d002119d036"} Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.466098 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwx7" Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.466119 4761 scope.go:117] "RemoveContainer" containerID="064c2716a216e98a952041bfc5ac72306c37a8ddf4cf000246f6b3117ad3cfff" Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.472828 4761 generic.go:334] "Generic (PLEG): container finished" podID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerID="f77b20dcf8c626671498098d027e29237b8f067dbd3887146360ba33013dca80" exitCode=0 Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.472867 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerDied","Data":"f77b20dcf8c626671498098d027e29237b8f067dbd3887146360ba33013dca80"} Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.490837 4761 scope.go:117] "RemoveContainer" containerID="0ead4f7bd5c80329344da831676280dfadd90f195d2e96dbe55ace1b64e71d03" Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.510763 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.512211 4761 scope.go:117] "RemoveContainer" containerID="1c350e067b76d01616548042ffe1b22dd7e6322fd5dff8554887ccdaadffa573" Nov 25 23:00:07 crc kubenswrapper[4761]: I1125 23:00:07.515904 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfwx7"] Nov 25 23:00:08 crc kubenswrapper[4761]: I1125 23:00:08.480825 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerStarted","Data":"de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818"} Nov 25 23:00:08 crc kubenswrapper[4761]: I1125 23:00:08.482525 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerStarted","Data":"324804955ddf37f7b9d77ab0edaced44e478c83ba0048525d21b019598326385"} Nov 25 23:00:08 crc kubenswrapper[4761]: I1125 23:00:08.512822 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gss6x" podStartSLOduration=3.174718092 podStartE2EDuration="56.512801252s" podCreationTimestamp="2025-11-25 22:59:12 +0000 UTC" firstStartedPulling="2025-11-25 22:59:14.915480196 +0000 UTC m=+150.648826031" lastFinishedPulling="2025-11-25 23:00:08.253563356 +0000 UTC m=+203.986909191" observedRunningTime="2025-11-25 23:00:08.508214074 +0000 UTC m=+204.241559939" watchObservedRunningTime="2025-11-25 23:00:08.512801252 +0000 UTC m=+204.246147087" Nov 25 23:00:08 crc kubenswrapper[4761]: I1125 23:00:08.527864 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4crxm" podStartSLOduration=2.652077854 podStartE2EDuration="53.527845024s" podCreationTimestamp="2025-11-25 22:59:15 +0000 UTC" firstStartedPulling="2025-11-25 22:59:17.023410367 +0000 UTC m=+152.756756202" lastFinishedPulling="2025-11-25 23:00:07.899177537 +0000 UTC m=+203.632523372" observedRunningTime="2025-11-25 23:00:08.525240908 +0000 UTC m=+204.258586763" watchObservedRunningTime="2025-11-25 23:00:08.527845024 +0000 UTC m=+204.261190859" Nov 25 23:00:09 crc kubenswrapper[4761]: I1125 23:00:09.020948 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" path="/var/lib/kubelet/pods/000888e9-b922-46cb-a11a-baf8ec7fa8f7/volumes" Nov 25 23:00:09 crc kubenswrapper[4761]: I1125 23:00:09.489493 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerStarted","Data":"da4c4b5f0847e1a0a69d32bb1f50103549ee240d12f93daac410bf9d1769cd66"} Nov 25 23:00:09 crc kubenswrapper[4761]: I1125 23:00:09.492167 4761 generic.go:334] "Generic (PLEG): container finished" podID="363875a6-dc39-4220-ab62-d8f390aa0625" containerID="4fe6aa41bdc86b3b13ca8f275614dc12ca6700fe7f4bb08d49e5ce85c69fa524" exitCode=0 Nov 25 23:00:09 crc kubenswrapper[4761]: I1125 23:00:09.492191 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerDied","Data":"4fe6aa41bdc86b3b13ca8f275614dc12ca6700fe7f4bb08d49e5ce85c69fa524"} Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.500486 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerStarted","Data":"81cebfafc795ff5c559807ea30e125c130ce6a3c7de666a78ceddc2ef0bb5f51"} Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.504105 4761 generic.go:334] "Generic (PLEG): container finished" podID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerID="f75078a2a6e64770afc69f1dd0786a828a6d4128ad33dce9b35595a0b41c7d0b" exitCode=0 Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.504171 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerDied","Data":"f75078a2a6e64770afc69f1dd0786a828a6d4128ad33dce9b35595a0b41c7d0b"} Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.507213 4761 generic.go:334] "Generic (PLEG): container finished" podID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerID="04da4f347015efc0ec17c80a503fb19b3e8250314f034ddcbbefbcdb43794007" exitCode=0 Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.507289 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerDied","Data":"04da4f347015efc0ec17c80a503fb19b3e8250314f034ddcbbefbcdb43794007"} Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.509191 4761 generic.go:334] "Generic (PLEG): container finished" podID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerID="da4c4b5f0847e1a0a69d32bb1f50103549ee240d12f93daac410bf9d1769cd66" exitCode=0 Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.509219 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerDied","Data":"da4c4b5f0847e1a0a69d32bb1f50103549ee240d12f93daac410bf9d1769cd66"} Nov 25 23:00:10 crc kubenswrapper[4761]: I1125 23:00:10.542730 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-56j56" podStartSLOduration=2.596412587 podStartE2EDuration="56.542694008s" podCreationTimestamp="2025-11-25 22:59:14 +0000 UTC" firstStartedPulling="2025-11-25 22:59:16.007856032 +0000 UTC m=+151.741201867" lastFinishedPulling="2025-11-25 23:00:09.954137453 +0000 UTC m=+205.687483288" observedRunningTime="2025-11-25 23:00:10.523615653 +0000 UTC m=+206.256961498" watchObservedRunningTime="2025-11-25 23:00:10.542694008 +0000 UTC m=+206.276039843" Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.518754 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerStarted","Data":"1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4"} Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.522110 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerStarted","Data":"0fe9f8ab217e8e8d4e5dc39f2606cdf9d452487de52232156b67ae3320627f8b"} Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.524818 4761 generic.go:334] "Generic (PLEG): container finished" podID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerID="e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a" exitCode=0 Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.524860 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerDied","Data":"e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a"} Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.527342 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerStarted","Data":"fcb75216a044d2be63d15985728699cb1a645bddff032e32c057490cd6361b8a"} Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.538659 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rdcmf" podStartSLOduration=2.387743184 podStartE2EDuration="59.538638893s" podCreationTimestamp="2025-11-25 22:59:12 +0000 UTC" firstStartedPulling="2025-11-25 22:59:13.799650281 +0000 UTC m=+149.532996116" lastFinishedPulling="2025-11-25 23:00:10.95054599 +0000 UTC m=+206.683891825" observedRunningTime="2025-11-25 23:00:11.534797725 +0000 UTC m=+207.268143570" watchObservedRunningTime="2025-11-25 23:00:11.538638893 +0000 UTC m=+207.271984728" Nov 25 23:00:11 crc kubenswrapper[4761]: I1125 23:00:11.551408 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-87tks" podStartSLOduration=2.7155952770000003 podStartE2EDuration="56.551390708s" podCreationTimestamp="2025-11-25 22:59:15 +0000 UTC" firstStartedPulling="2025-11-25 22:59:17.041319607 +0000 UTC m=+152.774665442" lastFinishedPulling="2025-11-25 23:00:10.877115038 +0000 UTC m=+206.610460873" observedRunningTime="2025-11-25 23:00:11.547804456 +0000 UTC m=+207.281150291" watchObservedRunningTime="2025-11-25 23:00:11.551390708 +0000 UTC m=+207.284736543" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.799730 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.800380 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.840705 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.860604 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9mf6" podStartSLOduration=3.950068346 podStartE2EDuration="57.860589003s" podCreationTimestamp="2025-11-25 22:59:15 +0000 UTC" firstStartedPulling="2025-11-25 22:59:17.017856008 +0000 UTC m=+152.751201843" lastFinishedPulling="2025-11-25 23:00:10.928376665 +0000 UTC m=+206.661722500" observedRunningTime="2025-11-25 23:00:11.593323066 +0000 UTC m=+207.326668911" watchObservedRunningTime="2025-11-25 23:00:12.860589003 +0000 UTC m=+208.593934838" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.924742 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.924793 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:00:12 crc kubenswrapper[4761]: I1125 23:00:12.968212 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:00:13 crc kubenswrapper[4761]: I1125 23:00:13.537987 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerStarted","Data":"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce"} Nov 25 23:00:13 crc kubenswrapper[4761]: I1125 23:00:13.553548 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9jt5l" podStartSLOduration=3.863542756 podStartE2EDuration="1m1.553533828s" podCreationTimestamp="2025-11-25 22:59:12 +0000 UTC" firstStartedPulling="2025-11-25 22:59:14.929679003 +0000 UTC m=+150.663024848" lastFinishedPulling="2025-11-25 23:00:12.619670085 +0000 UTC m=+208.353015920" observedRunningTime="2025-11-25 23:00:13.553125847 +0000 UTC m=+209.286471702" watchObservedRunningTime="2025-11-25 23:00:13.553533828 +0000 UTC m=+209.286879673" Nov 25 23:00:13 crc kubenswrapper[4761]: I1125 23:00:13.579617 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.115246 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.115550 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.157356 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.547256 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.547301 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.589472 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.600837 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.940560 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.940602 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:00:15 crc kubenswrapper[4761]: I1125 23:00:15.977606 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.191656 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.191767 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.247396 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.590727 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.595448 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:00:16 crc kubenswrapper[4761]: I1125 23:00:16.599017 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:19 crc kubenswrapper[4761]: I1125 23:00:19.358439 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 23:00:19 crc kubenswrapper[4761]: I1125 23:00:19.358991 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-87tks" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="registry-server" containerID="cri-o://0fe9f8ab217e8e8d4e5dc39f2606cdf9d452487de52232156b67ae3320627f8b" gracePeriod=2 Nov 25 23:00:20 crc kubenswrapper[4761]: I1125 23:00:20.763194 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 23:00:20 crc kubenswrapper[4761]: I1125 23:00:20.763534 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j9mf6" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="registry-server" containerID="cri-o://fcb75216a044d2be63d15985728699cb1a645bddff032e32c057490cd6361b8a" gracePeriod=2 Nov 25 23:00:21 crc kubenswrapper[4761]: I1125 23:00:21.128258 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:00:21 crc kubenswrapper[4761]: I1125 23:00:21.128653 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:00:21 crc kubenswrapper[4761]: I1125 23:00:21.128749 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:00:21 crc kubenswrapper[4761]: I1125 23:00:21.129615 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:00:21 crc kubenswrapper[4761]: I1125 23:00:21.129874 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68" gracePeriod=600 Nov 25 23:00:22 crc kubenswrapper[4761]: I1125 23:00:22.592823 4761 generic.go:334] "Generic (PLEG): container finished" podID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerID="0fe9f8ab217e8e8d4e5dc39f2606cdf9d452487de52232156b67ae3320627f8b" exitCode=0 Nov 25 23:00:22 crc kubenswrapper[4761]: I1125 23:00:22.592895 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerDied","Data":"0fe9f8ab217e8e8d4e5dc39f2606cdf9d452487de52232156b67ae3320627f8b"} Nov 25 23:00:22 crc kubenswrapper[4761]: I1125 23:00:22.874258 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.351569 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.356119 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.426049 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.609719 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68" exitCode=0 Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.609816 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68"} Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.613166 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87tks" event={"ID":"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2","Type":"ContainerDied","Data":"9196419038ba51312cc6d7f937bb86db7a1e02cb7bd420b8f6430aadb15b5d38"} Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.613211 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9196419038ba51312cc6d7f937bb86db7a1e02cb7bd420b8f6430aadb15b5d38" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.616388 4761 generic.go:334] "Generic (PLEG): container finished" podID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerID="fcb75216a044d2be63d15985728699cb1a645bddff032e32c057490cd6361b8a" exitCode=0 Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.616468 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerDied","Data":"fcb75216a044d2be63d15985728699cb1a645bddff032e32c057490cd6361b8a"} Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.656104 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.693954 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.740039 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28fjn\" (UniqueName: \"kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn\") pod \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.740159 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content\") pod \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.740211 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities\") pod \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\" (UID: \"8943ee05-ad49-4fe8-b4e2-104aac3fd2c2\") " Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.741966 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities" (OuterVolumeSpecName: "utilities") pod "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" (UID: "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.748943 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn" (OuterVolumeSpecName: "kube-api-access-28fjn") pod "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" (UID: "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2"). InnerVolumeSpecName "kube-api-access-28fjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.776773 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" (UID: "8943ee05-ad49-4fe8-b4e2-104aac3fd2c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.842536 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28fjn\" (UniqueName: \"kubernetes.io/projected/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-kube-api-access-28fjn\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.842594 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:23 crc kubenswrapper[4761]: I1125 23:00:23.842613 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.217110 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.348112 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content\") pod \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.348233 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities\") pod \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.348271 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r99hn\" (UniqueName: \"kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn\") pod \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\" (UID: \"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d\") " Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.349044 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities" (OuterVolumeSpecName: "utilities") pod "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" (UID: "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.352882 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn" (OuterVolumeSpecName: "kube-api-access-r99hn") pod "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" (UID: "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d"). InnerVolumeSpecName "kube-api-access-r99hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.449187 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.449222 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r99hn\" (UniqueName: \"kubernetes.io/projected/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-kube-api-access-r99hn\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.624745 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e"} Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.627302 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87tks" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.627592 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9mf6" event={"ID":"9b235fa2-d6a8-41b9-ab55-41fd4a60c90d","Type":"ContainerDied","Data":"c42631337247f8759f761dd3bbe28da23dceaa6645f75dcd77d4e2806fd42974"} Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.627638 4761 scope.go:117] "RemoveContainer" containerID="fcb75216a044d2be63d15985728699cb1a645bddff032e32c057490cd6361b8a" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.627760 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9mf6" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.641990 4761 scope.go:117] "RemoveContainer" containerID="da4c4b5f0847e1a0a69d32bb1f50103549ee240d12f93daac410bf9d1769cd66" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.668906 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.668954 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-87tks"] Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.682640 4761 scope.go:117] "RemoveContainer" containerID="21623d16205bbd6355c0661eb8dd3c24d0f523f6eb827640119a1907ff7eddfb" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.729360 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" (UID: "9b235fa2-d6a8-41b9-ab55-41fd4a60c90d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.753621 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.963455 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 23:00:24 crc kubenswrapper[4761]: I1125 23:00:24.967167 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j9mf6"] Nov 25 23:00:25 crc kubenswrapper[4761]: I1125 23:00:25.022915 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" path="/var/lib/kubelet/pods/8943ee05-ad49-4fe8-b4e2-104aac3fd2c2/volumes" Nov 25 23:00:25 crc kubenswrapper[4761]: I1125 23:00:25.023903 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" path="/var/lib/kubelet/pods/9b235fa2-d6a8-41b9-ab55-41fd4a60c90d/volumes" Nov 25 23:00:25 crc kubenswrapper[4761]: I1125 23:00:25.767357 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 23:00:25 crc kubenswrapper[4761]: I1125 23:00:25.767727 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9jt5l" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="registry-server" containerID="cri-o://72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce" gracePeriod=2 Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.117313 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.274279 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l4l8\" (UniqueName: \"kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8\") pod \"156832a0-c911-4ccf-bb41-511cbcdaab8c\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.274336 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities\") pod \"156832a0-c911-4ccf-bb41-511cbcdaab8c\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.274365 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content\") pod \"156832a0-c911-4ccf-bb41-511cbcdaab8c\" (UID: \"156832a0-c911-4ccf-bb41-511cbcdaab8c\") " Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.279557 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities" (OuterVolumeSpecName: "utilities") pod "156832a0-c911-4ccf-bb41-511cbcdaab8c" (UID: "156832a0-c911-4ccf-bb41-511cbcdaab8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.290877 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8" (OuterVolumeSpecName: "kube-api-access-5l4l8") pod "156832a0-c911-4ccf-bb41-511cbcdaab8c" (UID: "156832a0-c911-4ccf-bb41-511cbcdaab8c"). InnerVolumeSpecName "kube-api-access-5l4l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.326381 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "156832a0-c911-4ccf-bb41-511cbcdaab8c" (UID: "156832a0-c911-4ccf-bb41-511cbcdaab8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.375797 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l4l8\" (UniqueName: \"kubernetes.io/projected/156832a0-c911-4ccf-bb41-511cbcdaab8c-kube-api-access-5l4l8\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.375848 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.375862 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156832a0-c911-4ccf-bb41-511cbcdaab8c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.536781 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.641596 4761 generic.go:334] "Generic (PLEG): container finished" podID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerID="72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce" exitCode=0 Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.641866 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerDied","Data":"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce"} Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.641891 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jt5l" event={"ID":"156832a0-c911-4ccf-bb41-511cbcdaab8c","Type":"ContainerDied","Data":"78908cdd691aa51cf4fe0f8639cbaad8cb3a35254afbd960e5d77a062d88db8c"} Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.641907 4761 scope.go:117] "RemoveContainer" containerID="72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.642007 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jt5l" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.659237 4761 scope.go:117] "RemoveContainer" containerID="e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.668356 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.674420 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9jt5l"] Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.686748 4761 scope.go:117] "RemoveContainer" containerID="d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.699988 4761 scope.go:117] "RemoveContainer" containerID="72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce" Nov 25 23:00:26 crc kubenswrapper[4761]: E1125 23:00:26.700457 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce\": container with ID starting with 72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce not found: ID does not exist" containerID="72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.700487 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce"} err="failed to get container status \"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce\": rpc error: code = NotFound desc = could not find container \"72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce\": container with ID starting with 72da3590737646ad2468b59823b078a7447b669b12b1f87f4631b11d3f8c29ce not found: ID does not exist" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.700506 4761 scope.go:117] "RemoveContainer" containerID="e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a" Nov 25 23:00:26 crc kubenswrapper[4761]: E1125 23:00:26.700875 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a\": container with ID starting with e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a not found: ID does not exist" containerID="e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.700909 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a"} err="failed to get container status \"e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a\": rpc error: code = NotFound desc = could not find container \"e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a\": container with ID starting with e04e69df0d01148993f4e78d772097d64c61af4c0258b7a8c8c558edb063d65a not found: ID does not exist" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.700935 4761 scope.go:117] "RemoveContainer" containerID="d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c" Nov 25 23:00:26 crc kubenswrapper[4761]: E1125 23:00:26.701297 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c\": container with ID starting with d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c not found: ID does not exist" containerID="d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c" Nov 25 23:00:26 crc kubenswrapper[4761]: I1125 23:00:26.701338 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c"} err="failed to get container status \"d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c\": rpc error: code = NotFound desc = could not find container \"d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c\": container with ID starting with d4d32d69665f362eb0197b8d6513716a5d73a0f577650211d168e6ae4505eb9c not found: ID does not exist" Nov 25 23:00:27 crc kubenswrapper[4761]: I1125 23:00:27.020917 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" path="/var/lib/kubelet/pods/156832a0-c911-4ccf-bb41-511cbcdaab8c/volumes" Nov 25 23:00:51 crc kubenswrapper[4761]: I1125 23:00:51.572970 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerName="oauth-openshift" containerID="cri-o://1c8c2eb17156c0dae8c83aec1b981e65efe99c12caa2753d4f49c481589a353c" gracePeriod=15 Nov 25 23:00:51 crc kubenswrapper[4761]: I1125 23:00:51.817126 4761 generic.go:334] "Generic (PLEG): container finished" podID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerID="1c8c2eb17156c0dae8c83aec1b981e65efe99c12caa2753d4f49c481589a353c" exitCode=0 Nov 25 23:00:51 crc kubenswrapper[4761]: I1125 23:00:51.817385 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" event={"ID":"401d0f53-1c92-4d21-a8f8-8989bb51b0b9","Type":"ContainerDied","Data":"1c8c2eb17156c0dae8c83aec1b981e65efe99c12caa2753d4f49c481589a353c"} Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.003324 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037544 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-f94d7b7c5-mn74s"] Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037832 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037849 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037867 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerName="oauth-openshift" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037875 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerName="oauth-openshift" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037887 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037898 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037916 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037927 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037941 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037951 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037968 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037979 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.037991 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.037999 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="extract-content" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038012 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038021 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038038 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038047 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038060 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038068 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038079 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038088 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="extract-utilities" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038102 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038120 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038138 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038149 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: E1125 23:00:52.038162 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0984f952-54df-438b-929a-bb5ad66c2025" containerName="collect-profiles" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038172 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="0984f952-54df-438b-929a-bb5ad66c2025" containerName="collect-profiles" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038297 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="8943ee05-ad49-4fe8-b4e2-104aac3fd2c2" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038314 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b235fa2-d6a8-41b9-ab55-41fd4a60c90d" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038327 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="0984f952-54df-438b-929a-bb5ad66c2025" containerName="collect-profiles" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038342 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="156832a0-c911-4ccf-bb41-511cbcdaab8c" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038354 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" containerName="oauth-openshift" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038365 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="000888e9-b922-46cb-a11a-baf8ec7fa8f7" containerName="registry-server" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.038839 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.053157 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-f94d7b7c5-mn74s"] Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.136333 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.137930 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.137976 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.138015 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.138055 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.138094 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.138121 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.138162 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139017 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139112 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqvvd\" (UniqueName: \"kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139171 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139212 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139256 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139291 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139328 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139372 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca\") pod \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\" (UID: \"401d0f53-1c92-4d21-a8f8-8989bb51b0b9\") " Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139531 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139657 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-error\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139729 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139768 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-service-ca\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139873 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139916 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-session\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.139968 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140007 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140027 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140058 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140099 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-784gb\" (UniqueName: \"kubernetes.io/projected/0c185bd9-0646-46e8-8f68-5b893128e74d-kube-api-access-784gb\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140160 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-policies\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140205 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140559 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-dir\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140742 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-login\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140943 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-router-certs\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.140968 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.141092 4761 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.141117 4761 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.141138 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.141151 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.143067 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd" (OuterVolumeSpecName: "kube-api-access-kqvvd") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "kube-api-access-kqvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.144172 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.144587 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.144829 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.145571 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.147999 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.148211 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.152214 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.152435 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "401d0f53-1c92-4d21-a8f8-8989bb51b0b9" (UID: "401d0f53-1c92-4d21-a8f8-8989bb51b0b9"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242183 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242243 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-784gb\" (UniqueName: \"kubernetes.io/projected/0c185bd9-0646-46e8-8f68-5b893128e74d-kube-api-access-784gb\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242280 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-policies\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242318 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-dir\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242343 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-login\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242383 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-router-certs\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242408 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242431 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-error\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242454 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242479 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-service-ca\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242502 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242527 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-session\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242610 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242637 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242685 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242721 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqvvd\" (UniqueName: \"kubernetes.io/projected/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-kube-api-access-kqvvd\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242734 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242749 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242761 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242774 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242788 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242800 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242812 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.242825 4761 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/401d0f53-1c92-4d21-a8f8-8989bb51b0b9-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.243339 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-dir\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.244732 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-service-ca\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.244822 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-audit-policies\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.245061 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.245437 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.248539 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-router-certs\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.249784 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-login\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.249884 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-error\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.250297 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-session\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.251458 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.251503 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.251872 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.252118 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0c185bd9-0646-46e8-8f68-5b893128e74d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.265132 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-784gb\" (UniqueName: \"kubernetes.io/projected/0c185bd9-0646-46e8-8f68-5b893128e74d-kube-api-access-784gb\") pod \"oauth-openshift-f94d7b7c5-mn74s\" (UID: \"0c185bd9-0646-46e8-8f68-5b893128e74d\") " pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.362256 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.828741 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" event={"ID":"401d0f53-1c92-4d21-a8f8-8989bb51b0b9","Type":"ContainerDied","Data":"3d025499eaa640c96788cfca94e09553894cfe64c95f966310a83b8fb5ae1e1f"} Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.828822 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pp97f" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.829223 4761 scope.go:117] "RemoveContainer" containerID="1c8c2eb17156c0dae8c83aec1b981e65efe99c12caa2753d4f49c481589a353c" Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.856404 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-f94d7b7c5-mn74s"] Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.898392 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 23:00:52 crc kubenswrapper[4761]: I1125 23:00:52.907619 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pp97f"] Nov 25 23:00:53 crc kubenswrapper[4761]: I1125 23:00:53.023220 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401d0f53-1c92-4d21-a8f8-8989bb51b0b9" path="/var/lib/kubelet/pods/401d0f53-1c92-4d21-a8f8-8989bb51b0b9/volumes" Nov 25 23:00:53 crc kubenswrapper[4761]: I1125 23:00:53.836267 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" event={"ID":"0c185bd9-0646-46e8-8f68-5b893128e74d","Type":"ContainerStarted","Data":"739b95f7d1fdaa847b1ca18cbefcac10d590d15b5ad362c82295b74016d5fec4"} Nov 25 23:00:53 crc kubenswrapper[4761]: I1125 23:00:53.836741 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" event={"ID":"0c185bd9-0646-46e8-8f68-5b893128e74d","Type":"ContainerStarted","Data":"3bcee04882502de802d2c925aeb2a10ebc1e23a7e5d6d01d7abfedeb6f801bfc"} Nov 25 23:00:53 crc kubenswrapper[4761]: I1125 23:00:53.836776 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:00:53 crc kubenswrapper[4761]: I1125 23:00:53.866865 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" podStartSLOduration=27.866846722 podStartE2EDuration="27.866846722s" podCreationTimestamp="2025-11-25 23:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:00:53.863991009 +0000 UTC m=+249.597336864" watchObservedRunningTime="2025-11-25 23:00:53.866846722 +0000 UTC m=+249.600192557" Nov 25 23:00:54 crc kubenswrapper[4761]: I1125 23:00:54.029225 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-f94d7b7c5-mn74s" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.517540 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.518231 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rdcmf" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="registry-server" containerID="cri-o://1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" gracePeriod=30 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.524573 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.524800 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gss6x" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="registry-server" containerID="cri-o://de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" gracePeriod=30 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.540665 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.542238 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" containerID="cri-o://0368cb1288751616429b9f1059698c8f8e8c20b9e5a89cfda0fbeb12821bc4cb" gracePeriod=30 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.546818 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.547188 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-56j56" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="registry-server" containerID="cri-o://81cebfafc795ff5c559807ea30e125c130ce6a3c7de666a78ceddc2ef0bb5f51" gracePeriod=30 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.553867 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.554091 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4crxm" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="registry-server" containerID="cri-o://324804955ddf37f7b9d77ab0edaced44e478c83ba0048525d21b019598326385" gracePeriod=30 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.563663 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2hzp"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.566788 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.576390 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2hzp"] Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.724400 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngxs8\" (UniqueName: \"kubernetes.io/projected/ff22925e-febe-45b5-8cf7-7de17a24c7f1-kube-api-access-ngxs8\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.724706 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.724917 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.801483 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4 is running failed: container process not found" containerID="1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.802034 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4 is running failed: container process not found" containerID="1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.802565 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4 is running failed: container process not found" containerID="1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.802631 4761 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-rdcmf" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="registry-server" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.826459 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.826521 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngxs8\" (UniqueName: \"kubernetes.io/projected/ff22925e-febe-45b5-8cf7-7de17a24c7f1-kube-api-access-ngxs8\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.826546 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.830138 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.833378 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff22925e-febe-45b5-8cf7-7de17a24c7f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.847146 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngxs8\" (UniqueName: \"kubernetes.io/projected/ff22925e-febe-45b5-8cf7-7de17a24c7f1-kube-api-access-ngxs8\") pod \"marketplace-operator-79b997595-j2hzp\" (UID: \"ff22925e-febe-45b5-8cf7-7de17a24c7f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.881100 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.924886 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818 is running failed: container process not found" containerID="de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.925241 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818 is running failed: container process not found" containerID="de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.925512 4761 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818 is running failed: container process not found" containerID="de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 23:01:12 crc kubenswrapper[4761]: E1125 23:01:12.925533 4761 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-gss6x" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="registry-server" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.960810 4761 generic.go:334] "Generic (PLEG): container finished" podID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerID="1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" exitCode=0 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.960881 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerDied","Data":"1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.960907 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdcmf" event={"ID":"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a","Type":"ContainerDied","Data":"dbdf8ca9418dbeebeeeea1b17cc626d27f80cd27b7fc7920abaf409c24523d9f"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.960933 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbdf8ca9418dbeebeeeea1b17cc626d27f80cd27b7fc7920abaf409c24523d9f" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.963786 4761 generic.go:334] "Generic (PLEG): container finished" podID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerID="0368cb1288751616429b9f1059698c8f8e8c20b9e5a89cfda0fbeb12821bc4cb" exitCode=0 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.963856 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" event={"ID":"fd8e84c9-b40f-4e03-a91a-508fd653ee07","Type":"ContainerDied","Data":"0368cb1288751616429b9f1059698c8f8e8c20b9e5a89cfda0fbeb12821bc4cb"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.963884 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" event={"ID":"fd8e84c9-b40f-4e03-a91a-508fd653ee07","Type":"ContainerDied","Data":"bb07540777f25ef9946b9ce7193b39ddd1c7b72297220e8e032ad69253b5be91"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.963894 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb07540777f25ef9946b9ce7193b39ddd1c7b72297220e8e032ad69253b5be91" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.967846 4761 generic.go:334] "Generic (PLEG): container finished" podID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerID="de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" exitCode=0 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.967892 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerDied","Data":"de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.967912 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gss6x" event={"ID":"a310e19b-3b47-4c9c-ba97-9e4761d75fa9","Type":"ContainerDied","Data":"f33ab398643901f3bb54786db32d71e65fb672ba8c9d4471999762229045ec7f"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.967922 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f33ab398643901f3bb54786db32d71e65fb672ba8c9d4471999762229045ec7f" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.979062 4761 generic.go:334] "Generic (PLEG): container finished" podID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerID="324804955ddf37f7b9d77ab0edaced44e478c83ba0048525d21b019598326385" exitCode=0 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.979126 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerDied","Data":"324804955ddf37f7b9d77ab0edaced44e478c83ba0048525d21b019598326385"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.979184 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4crxm" event={"ID":"7a915718-fdf1-4829-9d76-a704fe9cd833","Type":"ContainerDied","Data":"866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.979196 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="866c9cecc08b3dad6744587a1453aa31c3a2ccb10b672235cdfaf3504f87627a" Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.985738 4761 generic.go:334] "Generic (PLEG): container finished" podID="363875a6-dc39-4220-ab62-d8f390aa0625" containerID="81cebfafc795ff5c559807ea30e125c130ce6a3c7de666a78ceddc2ef0bb5f51" exitCode=0 Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.985784 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerDied","Data":"81cebfafc795ff5c559807ea30e125c130ce6a3c7de666a78ceddc2ef0bb5f51"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.985825 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j56" event={"ID":"363875a6-dc39-4220-ab62-d8f390aa0625","Type":"ContainerDied","Data":"c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a"} Nov 25 23:01:12 crc kubenswrapper[4761]: I1125 23:01:12.985879 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1839a16f7e0ac63cde0d465c0ecc8642e1f547de107a38b6d8dec4c224cb68a" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.004503 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.016591 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.038353 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.052554 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.054223 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.093792 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2hzp"] Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131595 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca\") pod \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131647 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities\") pod \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131691 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics\") pod \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131744 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfk8p\" (UniqueName: \"kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p\") pod \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131761 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content\") pod \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\" (UID: \"33e7b5be-2cc4-4598-b197-e34a5c3d4e0a\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.131792 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gql82\" (UniqueName: \"kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82\") pod \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\" (UID: \"fd8e84c9-b40f-4e03-a91a-508fd653ee07\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.133131 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fd8e84c9-b40f-4e03-a91a-508fd653ee07" (UID: "fd8e84c9-b40f-4e03-a91a-508fd653ee07"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.133584 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities" (OuterVolumeSpecName: "utilities") pod "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" (UID: "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.137970 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fd8e84c9-b40f-4e03-a91a-508fd653ee07" (UID: "fd8e84c9-b40f-4e03-a91a-508fd653ee07"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.138429 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p" (OuterVolumeSpecName: "kube-api-access-dfk8p") pod "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" (UID: "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a"). InnerVolumeSpecName "kube-api-access-dfk8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.138465 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82" (OuterVolumeSpecName: "kube-api-access-gql82") pod "fd8e84c9-b40f-4e03-a91a-508fd653ee07" (UID: "fd8e84c9-b40f-4e03-a91a-508fd653ee07"). InnerVolumeSpecName "kube-api-access-gql82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.174983 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" (UID: "33e7b5be-2cc4-4598-b197-e34a5c3d4e0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234399 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content\") pod \"7a915718-fdf1-4829-9d76-a704fe9cd833\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234451 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities\") pod \"7a915718-fdf1-4829-9d76-a704fe9cd833\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234478 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h2l9\" (UniqueName: \"kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9\") pod \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234528 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content\") pod \"363875a6-dc39-4220-ab62-d8f390aa0625\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234582 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggvsx\" (UniqueName: \"kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx\") pod \"363875a6-dc39-4220-ab62-d8f390aa0625\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234609 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities\") pod \"363875a6-dc39-4220-ab62-d8f390aa0625\" (UID: \"363875a6-dc39-4220-ab62-d8f390aa0625\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234626 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities\") pod \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234644 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmf42\" (UniqueName: \"kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42\") pod \"7a915718-fdf1-4829-9d76-a704fe9cd833\" (UID: \"7a915718-fdf1-4829-9d76-a704fe9cd833\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234665 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content\") pod \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\" (UID: \"a310e19b-3b47-4c9c-ba97-9e4761d75fa9\") " Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234847 4761 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234863 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234873 4761 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd8e84c9-b40f-4e03-a91a-508fd653ee07-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234883 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfk8p\" (UniqueName: \"kubernetes.io/projected/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-kube-api-access-dfk8p\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234891 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.234900 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gql82\" (UniqueName: \"kubernetes.io/projected/fd8e84c9-b40f-4e03-a91a-508fd653ee07-kube-api-access-gql82\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.235881 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities" (OuterVolumeSpecName: "utilities") pod "7a915718-fdf1-4829-9d76-a704fe9cd833" (UID: "7a915718-fdf1-4829-9d76-a704fe9cd833"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.237279 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities" (OuterVolumeSpecName: "utilities") pod "363875a6-dc39-4220-ab62-d8f390aa0625" (UID: "363875a6-dc39-4220-ab62-d8f390aa0625"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.240016 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9" (OuterVolumeSpecName: "kube-api-access-9h2l9") pod "a310e19b-3b47-4c9c-ba97-9e4761d75fa9" (UID: "a310e19b-3b47-4c9c-ba97-9e4761d75fa9"). InnerVolumeSpecName "kube-api-access-9h2l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.241341 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx" (OuterVolumeSpecName: "kube-api-access-ggvsx") pod "363875a6-dc39-4220-ab62-d8f390aa0625" (UID: "363875a6-dc39-4220-ab62-d8f390aa0625"). InnerVolumeSpecName "kube-api-access-ggvsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.241561 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42" (OuterVolumeSpecName: "kube-api-access-cmf42") pod "7a915718-fdf1-4829-9d76-a704fe9cd833" (UID: "7a915718-fdf1-4829-9d76-a704fe9cd833"). InnerVolumeSpecName "kube-api-access-cmf42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.244313 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities" (OuterVolumeSpecName: "utilities") pod "a310e19b-3b47-4c9c-ba97-9e4761d75fa9" (UID: "a310e19b-3b47-4c9c-ba97-9e4761d75fa9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.261898 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "363875a6-dc39-4220-ab62-d8f390aa0625" (UID: "363875a6-dc39-4220-ab62-d8f390aa0625"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.295806 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a310e19b-3b47-4c9c-ba97-9e4761d75fa9" (UID: "a310e19b-3b47-4c9c-ba97-9e4761d75fa9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335681 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335729 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggvsx\" (UniqueName: \"kubernetes.io/projected/363875a6-dc39-4220-ab62-d8f390aa0625-kube-api-access-ggvsx\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335741 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363875a6-dc39-4220-ab62-d8f390aa0625-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335750 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335758 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmf42\" (UniqueName: \"kubernetes.io/projected/7a915718-fdf1-4829-9d76-a704fe9cd833-kube-api-access-cmf42\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335766 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335774 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.335781 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h2l9\" (UniqueName: \"kubernetes.io/projected/a310e19b-3b47-4c9c-ba97-9e4761d75fa9-kube-api-access-9h2l9\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.342686 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a915718-fdf1-4829-9d76-a704fe9cd833" (UID: "7a915718-fdf1-4829-9d76-a704fe9cd833"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.437402 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a915718-fdf1-4829-9d76-a704fe9cd833-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993151 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cqrtc" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993214 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gss6x" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993246 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdcmf" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993268 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j56" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993267 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4crxm" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.993283 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" event={"ID":"ff22925e-febe-45b5-8cf7-7de17a24c7f1","Type":"ContainerStarted","Data":"ddb0f90de1ed93f0d3a488695d5d813a2ff0b1591b2f1b0c34a49baabfb4d4c8"} Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.995564 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:13 crc kubenswrapper[4761]: I1125 23:01:13.995606 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" event={"ID":"ff22925e-febe-45b5-8cf7-7de17a24c7f1","Type":"ContainerStarted","Data":"ed7718db95d56817ace0cf6edbe507f9cdae830acc4363db40a7022e774a1155"} Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.004605 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.015562 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-j2hzp" podStartSLOduration=2.015546797 podStartE2EDuration="2.015546797s" podCreationTimestamp="2025-11-25 23:01:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:01:14.014951551 +0000 UTC m=+269.748297396" watchObservedRunningTime="2025-11-25 23:01:14.015546797 +0000 UTC m=+269.748892632" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.073833 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.080343 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cqrtc"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.087132 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.090259 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j56"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.098242 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.105293 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rdcmf"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.115436 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.124993 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gss6x"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.128471 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.131014 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4crxm"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735526 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vcncj"] Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735714 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735729 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735740 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735747 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735759 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735767 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735775 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735782 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735793 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735800 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735811 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735819 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735827 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735835 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735851 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735859 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735868 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735875 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="extract-utilities" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735885 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735892 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735900 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735907 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735918 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735924 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="extract-content" Nov 25 23:01:14 crc kubenswrapper[4761]: E1125 23:01:14.735931 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.735937 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736012 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736024 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736034 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736041 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" containerName="registry-server" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736049 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" containerName="marketplace-operator" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.736708 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.741144 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.780106 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vcncj"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.858679 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4swjk\" (UniqueName: \"kubernetes.io/projected/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-kube-api-access-4swjk\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.858774 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-utilities\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.858872 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-catalog-content\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.930357 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fnwm8"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.933336 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.935276 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.941727 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnwm8"] Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959273 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4swjk\" (UniqueName: \"kubernetes.io/projected/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-kube-api-access-4swjk\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959322 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-utilities\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959362 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx5nn\" (UniqueName: \"kubernetes.io/projected/534dc1b1-03d3-4b67-bbdc-a4ca72556281-kube-api-access-gx5nn\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959390 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-catalog-content\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959423 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-catalog-content\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.959455 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-utilities\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.960310 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-utilities\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.960316 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-catalog-content\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:14 crc kubenswrapper[4761]: I1125 23:01:14.985787 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4swjk\" (UniqueName: \"kubernetes.io/projected/9aacd34f-6d5f-4a1a-8a47-242df1d020d7-kube-api-access-4swjk\") pod \"community-operators-vcncj\" (UID: \"9aacd34f-6d5f-4a1a-8a47-242df1d020d7\") " pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.018794 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e7b5be-2cc4-4598-b197-e34a5c3d4e0a" path="/var/lib/kubelet/pods/33e7b5be-2cc4-4598-b197-e34a5c3d4e0a/volumes" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.020111 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363875a6-dc39-4220-ab62-d8f390aa0625" path="/var/lib/kubelet/pods/363875a6-dc39-4220-ab62-d8f390aa0625/volumes" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.020943 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a915718-fdf1-4829-9d76-a704fe9cd833" path="/var/lib/kubelet/pods/7a915718-fdf1-4829-9d76-a704fe9cd833/volumes" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.022137 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a310e19b-3b47-4c9c-ba97-9e4761d75fa9" path="/var/lib/kubelet/pods/a310e19b-3b47-4c9c-ba97-9e4761d75fa9/volumes" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.022781 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8e84c9-b40f-4e03-a91a-508fd653ee07" path="/var/lib/kubelet/pods/fd8e84c9-b40f-4e03-a91a-508fd653ee07/volumes" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.056031 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.060011 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-utilities\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.060167 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx5nn\" (UniqueName: \"kubernetes.io/projected/534dc1b1-03d3-4b67-bbdc-a4ca72556281-kube-api-access-gx5nn\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.060259 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-catalog-content\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.060533 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-utilities\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.061149 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534dc1b1-03d3-4b67-bbdc-a4ca72556281-catalog-content\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.078649 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx5nn\" (UniqueName: \"kubernetes.io/projected/534dc1b1-03d3-4b67-bbdc-a4ca72556281-kube-api-access-gx5nn\") pod \"certified-operators-fnwm8\" (UID: \"534dc1b1-03d3-4b67-bbdc-a4ca72556281\") " pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.253975 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.286989 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vcncj"] Nov 25 23:01:15 crc kubenswrapper[4761]: W1125 23:01:15.300768 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aacd34f_6d5f_4a1a_8a47_242df1d020d7.slice/crio-9e914445439aeca1b4f8b57a358598fdc3dbddb76228a28cd9e18de9621d1e5f WatchSource:0}: Error finding container 9e914445439aeca1b4f8b57a358598fdc3dbddb76228a28cd9e18de9621d1e5f: Status 404 returned error can't find the container with id 9e914445439aeca1b4f8b57a358598fdc3dbddb76228a28cd9e18de9621d1e5f Nov 25 23:01:15 crc kubenswrapper[4761]: I1125 23:01:15.419436 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnwm8"] Nov 25 23:01:15 crc kubenswrapper[4761]: W1125 23:01:15.425444 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod534dc1b1_03d3_4b67_bbdc_a4ca72556281.slice/crio-3dcd57d3ed62925c246d7f3f352fa9aff184d87d54ffdee579d398d0cf37cff0 WatchSource:0}: Error finding container 3dcd57d3ed62925c246d7f3f352fa9aff184d87d54ffdee579d398d0cf37cff0: Status 404 returned error can't find the container with id 3dcd57d3ed62925c246d7f3f352fa9aff184d87d54ffdee579d398d0cf37cff0 Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.004875 4761 generic.go:334] "Generic (PLEG): container finished" podID="9aacd34f-6d5f-4a1a-8a47-242df1d020d7" containerID="e1182f9cdb34d5e0eceb79e7153a5cbe324508d70a26686d7264c346b9b81907" exitCode=0 Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.004932 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcncj" event={"ID":"9aacd34f-6d5f-4a1a-8a47-242df1d020d7","Type":"ContainerDied","Data":"e1182f9cdb34d5e0eceb79e7153a5cbe324508d70a26686d7264c346b9b81907"} Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.005289 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcncj" event={"ID":"9aacd34f-6d5f-4a1a-8a47-242df1d020d7","Type":"ContainerStarted","Data":"9e914445439aeca1b4f8b57a358598fdc3dbddb76228a28cd9e18de9621d1e5f"} Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.007819 4761 generic.go:334] "Generic (PLEG): container finished" podID="534dc1b1-03d3-4b67-bbdc-a4ca72556281" containerID="0230d5e2025cbac05a577fa678b00311839382ce0948ea4f1589190f5ea7c582" exitCode=0 Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.008494 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwm8" event={"ID":"534dc1b1-03d3-4b67-bbdc-a4ca72556281","Type":"ContainerDied","Data":"0230d5e2025cbac05a577fa678b00311839382ce0948ea4f1589190f5ea7c582"} Nov 25 23:01:16 crc kubenswrapper[4761]: I1125 23:01:16.008546 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwm8" event={"ID":"534dc1b1-03d3-4b67-bbdc-a4ca72556281","Type":"ContainerStarted","Data":"3dcd57d3ed62925c246d7f3f352fa9aff184d87d54ffdee579d398d0cf37cff0"} Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.016852 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwm8" event={"ID":"534dc1b1-03d3-4b67-bbdc-a4ca72556281","Type":"ContainerStarted","Data":"974862cd292da9939c004e3d08eba0f8ca0a31b5f52d829c88f45375a6c55ae2"} Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.018296 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcncj" event={"ID":"9aacd34f-6d5f-4a1a-8a47-242df1d020d7","Type":"ContainerDied","Data":"78c917f1fca93429d241c3907e8bab09ab842b9feee89f642d65b5a755e0b28f"} Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.017784 4761 generic.go:334] "Generic (PLEG): container finished" podID="9aacd34f-6d5f-4a1a-8a47-242df1d020d7" containerID="78c917f1fca93429d241c3907e8bab09ab842b9feee89f642d65b5a755e0b28f" exitCode=0 Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.131560 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n9z85"] Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.132749 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.134275 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.162887 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9z85"] Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.213849 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj6j4\" (UniqueName: \"kubernetes.io/projected/23e39177-070d-4248-a455-7b2c98fd9486-kube-api-access-dj6j4\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.214044 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-utilities\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.214189 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-catalog-content\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.316024 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-catalog-content\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.316088 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj6j4\" (UniqueName: \"kubernetes.io/projected/23e39177-070d-4248-a455-7b2c98fd9486-kube-api-access-dj6j4\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.316114 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-utilities\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.316606 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-utilities\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.316922 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e39177-070d-4248-a455-7b2c98fd9486-catalog-content\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.337670 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h9c26"] Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.338624 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.342600 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.347089 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h9c26"] Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.347320 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj6j4\" (UniqueName: \"kubernetes.io/projected/23e39177-070d-4248-a455-7b2c98fd9486-kube-api-access-dj6j4\") pod \"redhat-marketplace-n9z85\" (UID: \"23e39177-070d-4248-a455-7b2c98fd9486\") " pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.437662 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcjh2\" (UniqueName: \"kubernetes.io/projected/9e4155b4-dab7-4344-a1f7-cb95429b20ca-kube-api-access-mcjh2\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.437819 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-utilities\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.437846 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-catalog-content\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.450711 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.539111 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcjh2\" (UniqueName: \"kubernetes.io/projected/9e4155b4-dab7-4344-a1f7-cb95429b20ca-kube-api-access-mcjh2\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.539164 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-utilities\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.539185 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-catalog-content\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.539634 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-catalog-content\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.539723 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e4155b4-dab7-4344-a1f7-cb95429b20ca-utilities\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.556062 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcjh2\" (UniqueName: \"kubernetes.io/projected/9e4155b4-dab7-4344-a1f7-cb95429b20ca-kube-api-access-mcjh2\") pod \"redhat-operators-h9c26\" (UID: \"9e4155b4-dab7-4344-a1f7-cb95429b20ca\") " pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.670833 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9z85"] Nov 25 23:01:17 crc kubenswrapper[4761]: W1125 23:01:17.681282 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23e39177_070d_4248_a455_7b2c98fd9486.slice/crio-059026d1669fdcacb4c12aa0bc2e12f750aaf4752a6adb38782b29f7cda4113a WatchSource:0}: Error finding container 059026d1669fdcacb4c12aa0bc2e12f750aaf4752a6adb38782b29f7cda4113a: Status 404 returned error can't find the container with id 059026d1669fdcacb4c12aa0bc2e12f750aaf4752a6adb38782b29f7cda4113a Nov 25 23:01:17 crc kubenswrapper[4761]: I1125 23:01:17.824031 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.017678 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h9c26"] Nov 25 23:01:18 crc kubenswrapper[4761]: W1125 23:01:18.025815 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e4155b4_dab7_4344_a1f7_cb95429b20ca.slice/crio-59be3ec17977644f3bd2bc4a86dc28b990f2fdb4bf0fcc78a5230ffb7c7bd18b WatchSource:0}: Error finding container 59be3ec17977644f3bd2bc4a86dc28b990f2fdb4bf0fcc78a5230ffb7c7bd18b: Status 404 returned error can't find the container with id 59be3ec17977644f3bd2bc4a86dc28b990f2fdb4bf0fcc78a5230ffb7c7bd18b Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.028659 4761 generic.go:334] "Generic (PLEG): container finished" podID="534dc1b1-03d3-4b67-bbdc-a4ca72556281" containerID="974862cd292da9939c004e3d08eba0f8ca0a31b5f52d829c88f45375a6c55ae2" exitCode=0 Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.028765 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwm8" event={"ID":"534dc1b1-03d3-4b67-bbdc-a4ca72556281","Type":"ContainerDied","Data":"974862cd292da9939c004e3d08eba0f8ca0a31b5f52d829c88f45375a6c55ae2"} Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.049664 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcncj" event={"ID":"9aacd34f-6d5f-4a1a-8a47-242df1d020d7","Type":"ContainerStarted","Data":"ef0bb20b040619f446bc867ca6a775768a639a87f7e7767fb032e1b9cd9390c9"} Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.053037 4761 generic.go:334] "Generic (PLEG): container finished" podID="23e39177-070d-4248-a455-7b2c98fd9486" containerID="44ea2ef96f7595a5396a0353f1ce06ea1024b653799e45968951fbddcee1984f" exitCode=0 Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.053077 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9z85" event={"ID":"23e39177-070d-4248-a455-7b2c98fd9486","Type":"ContainerDied","Data":"44ea2ef96f7595a5396a0353f1ce06ea1024b653799e45968951fbddcee1984f"} Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.053099 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9z85" event={"ID":"23e39177-070d-4248-a455-7b2c98fd9486","Type":"ContainerStarted","Data":"059026d1669fdcacb4c12aa0bc2e12f750aaf4752a6adb38782b29f7cda4113a"} Nov 25 23:01:18 crc kubenswrapper[4761]: I1125 23:01:18.096067 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vcncj" podStartSLOduration=2.633904597 podStartE2EDuration="4.096046911s" podCreationTimestamp="2025-11-25 23:01:14 +0000 UTC" firstStartedPulling="2025-11-25 23:01:16.006763503 +0000 UTC m=+271.740109378" lastFinishedPulling="2025-11-25 23:01:17.468905857 +0000 UTC m=+273.202251692" observedRunningTime="2025-11-25 23:01:18.075856227 +0000 UTC m=+273.809202072" watchObservedRunningTime="2025-11-25 23:01:18.096046911 +0000 UTC m=+273.829392756" Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.059856 4761 generic.go:334] "Generic (PLEG): container finished" podID="9e4155b4-dab7-4344-a1f7-cb95429b20ca" containerID="d4709161c13534693c9b9e8eacf112a543c2283c0ced0f1419467142531197c7" exitCode=0 Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.059988 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h9c26" event={"ID":"9e4155b4-dab7-4344-a1f7-cb95429b20ca","Type":"ContainerDied","Data":"d4709161c13534693c9b9e8eacf112a543c2283c0ced0f1419467142531197c7"} Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.060261 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h9c26" event={"ID":"9e4155b4-dab7-4344-a1f7-cb95429b20ca","Type":"ContainerStarted","Data":"59be3ec17977644f3bd2bc4a86dc28b990f2fdb4bf0fcc78a5230ffb7c7bd18b"} Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.062595 4761 generic.go:334] "Generic (PLEG): container finished" podID="23e39177-070d-4248-a455-7b2c98fd9486" containerID="2e52614b7a5815a9aaf1b75ccce235d94b7a130a6403b6500acd504d19318d1a" exitCode=0 Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.062617 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9z85" event={"ID":"23e39177-070d-4248-a455-7b2c98fd9486","Type":"ContainerDied","Data":"2e52614b7a5815a9aaf1b75ccce235d94b7a130a6403b6500acd504d19318d1a"} Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.066118 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwm8" event={"ID":"534dc1b1-03d3-4b67-bbdc-a4ca72556281","Type":"ContainerStarted","Data":"44531b5570f9055fa9e776ec866fa2a6b42b0395643e9bfd2ff26b28297200d5"} Nov 25 23:01:19 crc kubenswrapper[4761]: I1125 23:01:19.091657 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fnwm8" podStartSLOduration=2.386148215 podStartE2EDuration="5.091638034s" podCreationTimestamp="2025-11-25 23:01:14 +0000 UTC" firstStartedPulling="2025-11-25 23:01:16.011996718 +0000 UTC m=+271.745342593" lastFinishedPulling="2025-11-25 23:01:18.717486577 +0000 UTC m=+274.450832412" observedRunningTime="2025-11-25 23:01:19.090837332 +0000 UTC m=+274.824183187" watchObservedRunningTime="2025-11-25 23:01:19.091638034 +0000 UTC m=+274.824983869" Nov 25 23:01:20 crc kubenswrapper[4761]: I1125 23:01:20.076333 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h9c26" event={"ID":"9e4155b4-dab7-4344-a1f7-cb95429b20ca","Type":"ContainerStarted","Data":"07240abfae9081e0add4814feace46d5c8c21f8ac00e94b076411ee47dc735b8"} Nov 25 23:01:20 crc kubenswrapper[4761]: I1125 23:01:20.078715 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9z85" event={"ID":"23e39177-070d-4248-a455-7b2c98fd9486","Type":"ContainerStarted","Data":"ffcc94ed160ae1dbcecfe33f20fe11edf8ddb85ef94d9be016af1496d3efaead"} Nov 25 23:01:20 crc kubenswrapper[4761]: I1125 23:01:20.115160 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n9z85" podStartSLOduration=1.555695244 podStartE2EDuration="3.11514005s" podCreationTimestamp="2025-11-25 23:01:17 +0000 UTC" firstStartedPulling="2025-11-25 23:01:18.054224707 +0000 UTC m=+273.787570542" lastFinishedPulling="2025-11-25 23:01:19.613669503 +0000 UTC m=+275.347015348" observedRunningTime="2025-11-25 23:01:20.112799449 +0000 UTC m=+275.846145324" watchObservedRunningTime="2025-11-25 23:01:20.11514005 +0000 UTC m=+275.848485895" Nov 25 23:01:21 crc kubenswrapper[4761]: I1125 23:01:21.085523 4761 generic.go:334] "Generic (PLEG): container finished" podID="9e4155b4-dab7-4344-a1f7-cb95429b20ca" containerID="07240abfae9081e0add4814feace46d5c8c21f8ac00e94b076411ee47dc735b8" exitCode=0 Nov 25 23:01:21 crc kubenswrapper[4761]: I1125 23:01:21.085581 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h9c26" event={"ID":"9e4155b4-dab7-4344-a1f7-cb95429b20ca","Type":"ContainerDied","Data":"07240abfae9081e0add4814feace46d5c8c21f8ac00e94b076411ee47dc735b8"} Nov 25 23:01:23 crc kubenswrapper[4761]: I1125 23:01:23.100804 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h9c26" event={"ID":"9e4155b4-dab7-4344-a1f7-cb95429b20ca","Type":"ContainerStarted","Data":"e6d5318e048d60b3ba67dfcc78f40b1d3faab6d3f0d575a7847695a773d4d24b"} Nov 25 23:01:23 crc kubenswrapper[4761]: I1125 23:01:23.124290 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h9c26" podStartSLOduration=3.666943652 podStartE2EDuration="6.124263988s" podCreationTimestamp="2025-11-25 23:01:17 +0000 UTC" firstStartedPulling="2025-11-25 23:01:19.062081658 +0000 UTC m=+274.795427493" lastFinishedPulling="2025-11-25 23:01:21.519401994 +0000 UTC m=+277.252747829" observedRunningTime="2025-11-25 23:01:23.117882123 +0000 UTC m=+278.851227968" watchObservedRunningTime="2025-11-25 23:01:23.124263988 +0000 UTC m=+278.857609833" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.056204 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.056245 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.124620 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.193637 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vcncj" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.254630 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.254742 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:25 crc kubenswrapper[4761]: I1125 23:01:25.299250 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:26 crc kubenswrapper[4761]: I1125 23:01:26.166119 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fnwm8" Nov 25 23:01:27 crc kubenswrapper[4761]: I1125 23:01:27.452213 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:27 crc kubenswrapper[4761]: I1125 23:01:27.452674 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:27 crc kubenswrapper[4761]: I1125 23:01:27.514502 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:27 crc kubenswrapper[4761]: I1125 23:01:27.824524 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:27 crc kubenswrapper[4761]: I1125 23:01:27.824657 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:28 crc kubenswrapper[4761]: I1125 23:01:28.171542 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n9z85" Nov 25 23:01:28 crc kubenswrapper[4761]: I1125 23:01:28.913584 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h9c26" podUID="9e4155b4-dab7-4344-a1f7-cb95429b20ca" containerName="registry-server" probeResult="failure" output=< Nov 25 23:01:28 crc kubenswrapper[4761]: timeout: failed to connect service ":50051" within 1s Nov 25 23:01:28 crc kubenswrapper[4761]: > Nov 25 23:01:37 crc kubenswrapper[4761]: I1125 23:01:37.892401 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:01:37 crc kubenswrapper[4761]: I1125 23:01:37.996377 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h9c26" Nov 25 23:02:51 crc kubenswrapper[4761]: I1125 23:02:51.127723 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:02:51 crc kubenswrapper[4761]: I1125 23:02:51.128544 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:03:21 crc kubenswrapper[4761]: I1125 23:03:21.127627 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:03:21 crc kubenswrapper[4761]: I1125 23:03:21.128485 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.127777 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.128556 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.128627 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.129478 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.129579 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e" gracePeriod=600 Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.302200 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e" exitCode=0 Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.302298 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e"} Nov 25 23:03:51 crc kubenswrapper[4761]: I1125 23:03:51.302356 4761 scope.go:117] "RemoveContainer" containerID="6ed9deb75b2cc5710f9f14a09c7478e7203f36fcf0b895aa5f372bd904f1cf68" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.068449 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hgbwz"] Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.069953 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.084688 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hgbwz"] Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191454 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-tls\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191641 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191741 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-trusted-ca\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191778 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191920 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt4v4\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-kube-api-access-xt4v4\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.191957 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.192000 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-certificates\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.192041 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-bound-sa-token\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.220040 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.293729 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-certificates\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.293816 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-bound-sa-token\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.293903 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-tls\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.293938 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-trusted-ca\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.293973 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.294036 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt4v4\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-kube-api-access-xt4v4\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.294074 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.296184 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.296895 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-trusted-ca\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.297038 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-certificates\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.311135 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.311414 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-registry-tls\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.314616 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c"} Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.316354 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-bound-sa-token\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.322351 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt4v4\" (UniqueName: \"kubernetes.io/projected/275e7eaf-462a-4dcb-a37c-0c4323dd3d83-kube-api-access-xt4v4\") pod \"image-registry-66df7c8f76-hgbwz\" (UID: \"275e7eaf-462a-4dcb-a37c-0c4323dd3d83\") " pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.420039 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:03:52 crc kubenswrapper[4761]: I1125 23:03:52.681148 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hgbwz"] Nov 25 23:03:53 crc kubenswrapper[4761]: I1125 23:03:53.325676 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" event={"ID":"275e7eaf-462a-4dcb-a37c-0c4323dd3d83","Type":"ContainerStarted","Data":"4c613a24e134a44bb135caf671c153706be4d9d63891d7c977bec0c5610e17b6"} Nov 25 23:03:53 crc kubenswrapper[4761]: I1125 23:03:53.326132 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" event={"ID":"275e7eaf-462a-4dcb-a37c-0c4323dd3d83","Type":"ContainerStarted","Data":"c8c86dc1821404cc92d96856c17bd8ff6f525341099a5ec1193f69e074a9d36a"} Nov 25 23:03:53 crc kubenswrapper[4761]: I1125 23:03:53.351771 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" podStartSLOduration=1.351749801 podStartE2EDuration="1.351749801s" podCreationTimestamp="2025-11-25 23:03:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:03:53.347780255 +0000 UTC m=+429.081126120" watchObservedRunningTime="2025-11-25 23:03:53.351749801 +0000 UTC m=+429.085095646" Nov 25 23:03:54 crc kubenswrapper[4761]: I1125 23:03:54.341416 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:04:12 crc kubenswrapper[4761]: I1125 23:04:12.429602 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hgbwz" Nov 25 23:04:12 crc kubenswrapper[4761]: I1125 23:04:12.522791 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 23:04:37 crc kubenswrapper[4761]: I1125 23:04:37.581577 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" podUID="c80d2ace-7b69-4960-a664-41fb94d072de" containerName="registry" containerID="cri-o://c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427" gracePeriod=30 Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.021247 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056450 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056753 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056810 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfn49\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056841 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056876 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056899 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056917 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.056948 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted\") pod \"c80d2ace-7b69-4960-a664-41fb94d072de\" (UID: \"c80d2ace-7b69-4960-a664-41fb94d072de\") " Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.068072 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.068835 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.070198 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.076138 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.076871 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49" (OuterVolumeSpecName: "kube-api-access-tfn49") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "kube-api-access-tfn49". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.077420 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.085312 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.086795 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c80d2ace-7b69-4960-a664-41fb94d072de" (UID: "c80d2ace-7b69-4960-a664-41fb94d072de"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.127243 4761 generic.go:334] "Generic (PLEG): container finished" podID="c80d2ace-7b69-4960-a664-41fb94d072de" containerID="c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427" exitCode=0 Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.127293 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" event={"ID":"c80d2ace-7b69-4960-a664-41fb94d072de","Type":"ContainerDied","Data":"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427"} Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.127320 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" event={"ID":"c80d2ace-7b69-4960-a664-41fb94d072de","Type":"ContainerDied","Data":"c9824be5f6dfd9a447c2c47a0ae058527af245c0b42ff13d8615c71f7589782f"} Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.127337 4761 scope.go:117] "RemoveContainer" containerID="c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.127388 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t6smk" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.155508 4761 scope.go:117] "RemoveContainer" containerID="c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427" Nov 25 23:04:38 crc kubenswrapper[4761]: E1125 23:04:38.156080 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427\": container with ID starting with c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427 not found: ID does not exist" containerID="c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.156122 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427"} err="failed to get container status \"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427\": rpc error: code = NotFound desc = could not find container \"c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427\": container with ID starting with c72ef16607d23650a01fa41124dd729ed32b0bbbd13e059c7f2afb85b5a53427 not found: ID does not exist" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160323 4761 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160386 4761 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160402 4761 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c80d2ace-7b69-4960-a664-41fb94d072de-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160417 4761 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c80d2ace-7b69-4960-a664-41fb94d072de-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160461 4761 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c80d2ace-7b69-4960-a664-41fb94d072de-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160472 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfn49\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-kube-api-access-tfn49\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.160483 4761 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c80d2ace-7b69-4960-a664-41fb94d072de-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.167580 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 23:04:38 crc kubenswrapper[4761]: I1125 23:04:38.170244 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t6smk"] Nov 25 23:04:39 crc kubenswrapper[4761]: I1125 23:04:39.025547 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c80d2ace-7b69-4960-a664-41fb94d072de" path="/var/lib/kubelet/pods/c80d2ace-7b69-4960-a664-41fb94d072de/volumes" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.178267 4761 scope.go:117] "RemoveContainer" containerID="70ed616a2f61b8c050f1064a8506846cacfeb9a3ca2ddb7fc8ce0dc8d9411f1f" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.211297 4761 scope.go:117] "RemoveContainer" containerID="0368cb1288751616429b9f1059698c8f8e8c20b9e5a89cfda0fbeb12821bc4cb" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.250352 4761 scope.go:117] "RemoveContainer" containerID="1b14595e2051e6e9317c43cd6ff8719f252a761f9e467e31376252c544d639c0" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.264351 4761 scope.go:117] "RemoveContainer" containerID="30af171f5ea6ccf5e281bf19a85b0e3d31109da1ab7699711865102684d0b6e6" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.289384 4761 scope.go:117] "RemoveContainer" containerID="10b195a7b95dddb251cec4757fc93f79f88c56cc4ac0ceea4cce735772949a20" Nov 25 23:05:45 crc kubenswrapper[4761]: I1125 23:05:45.309339 4761 scope.go:117] "RemoveContainer" containerID="629acc97814673915258cd5a4f750e64e214bda7dcd3e94a543f9a3c0fbe824a" Nov 25 23:05:51 crc kubenswrapper[4761]: I1125 23:05:51.128153 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:05:51 crc kubenswrapper[4761]: I1125 23:05:51.128568 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:06:21 crc kubenswrapper[4761]: I1125 23:06:21.128396 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:06:21 crc kubenswrapper[4761]: I1125 23:06:21.129079 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.791590 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qdk9h"] Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.792860 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-controller" containerID="cri-o://f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.792929 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="sbdb" containerID="cri-o://e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.792939 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="nbdb" containerID="cri-o://c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.793074 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.793125 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="northd" containerID="cri-o://6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.793168 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-node" containerID="cri-o://507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.793207 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-acl-logging" containerID="cri-o://b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.826963 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" containerID="cri-o://76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" gracePeriod=30 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.972342 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovnkube-controller/3.log" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.974500 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-acl-logging/0.log" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975194 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-controller/0.log" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975572 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" exitCode=0 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975598 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" exitCode=0 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975607 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" exitCode=0 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975620 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" exitCode=143 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975601 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975658 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975674 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975688 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975712 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975630 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" exitCode=143 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.975731 4761 scope.go:117] "RemoveContainer" containerID="c47e4cd9096e149a2792e806c7b5109fc4a17f1b210bd1c9b7f1cdee28108176" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.977869 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/2.log" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.978289 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/1.log" Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.978320 4761 generic.go:334] "Generic (PLEG): container finished" podID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" containerID="438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4" exitCode=2 Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.978337 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerDied","Data":"438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4"} Nov 25 23:06:29 crc kubenswrapper[4761]: I1125 23:06:29.978743 4761 scope.go:117] "RemoveContainer" containerID="438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4" Nov 25 23:06:29 crc kubenswrapper[4761]: E1125 23:06:29.978895 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-fnqp9_openshift-multus(5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6)\"" pod="openshift-multus/multus-fnqp9" podUID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.033562 4761 scope.go:117] "RemoveContainer" containerID="aabb13b93104af9614a480a2654b3a9a50f8331ca99c0aa5786420d64b007577" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.161387 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-acl-logging/0.log" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.162419 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-controller/0.log" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.163184 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216433 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-czxbc"] Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216636 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kubecfg-setup" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216649 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kubecfg-setup" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216663 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="northd" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216674 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="northd" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216684 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216692 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216728 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216736 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216747 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216754 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216762 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-node" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216769 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-node" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216779 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="sbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216787 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="sbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216799 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216806 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216817 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80d2ace-7b69-4960-a664-41fb94d072de" containerName="registry" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216824 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80d2ace-7b69-4960-a664-41fb94d072de" containerName="registry" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216834 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216841 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216852 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216860 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216871 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-acl-logging" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216879 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-acl-logging" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.216893 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="nbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.216900 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="nbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217001 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217011 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="sbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217022 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217031 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-acl-logging" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217039 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="nbdb" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217049 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217060 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="c80d2ace-7b69-4960-a664-41fb94d072de" containerName="registry" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217071 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovn-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217082 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217094 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217101 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="northd" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217114 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="kube-rbac-proxy-node" Nov 25 23:06:30 crc kubenswrapper[4761]: E1125 23:06:30.217218 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217227 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.217327 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerName="ovnkube-controller" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.219170 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.341976 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342287 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342392 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342476 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342565 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342653 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342842 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342930 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343053 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343220 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343308 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343402 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lj29\" (UniqueName: \"kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343483 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343570 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343651 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343756 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.344360 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342405 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342534 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash" (OuterVolumeSpecName: "host-slash") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342607 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342788 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342849 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.342919 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343008 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343100 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343396 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.343870 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket" (OuterVolumeSpecName: "log-socket") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.344084 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.344101 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345270 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.344419 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.344444 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log" (OuterVolumeSpecName: "node-log") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345056 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345403 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345477 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin\") pod \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\" (UID: \"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8\") " Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345559 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345617 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345665 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/005e52a3-41d1-438f-b439-f68b6585e596-ovn-node-metrics-cert\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345770 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-kubelet\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.345924 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-netd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346069 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-netns\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346134 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-script-lib\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346162 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-var-lib-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346185 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-systemd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346253 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-node-log\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346354 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-ovn\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346413 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzwpw\" (UniqueName: \"kubernetes.io/projected/005e52a3-41d1-438f-b439-f68b6585e596-kube-api-access-qzwpw\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346457 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-etc-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346498 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346554 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346591 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-systemd-units\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346627 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346663 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-slash\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346751 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-config\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346804 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-bin\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346848 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-log-socket\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346900 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-env-overrides\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.346970 4761 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347090 4761 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347178 4761 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347251 4761 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347320 4761 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347389 4761 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347458 4761 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347525 4761 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347598 4761 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347667 4761 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347785 4761 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347860 4761 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.347929 4761 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.348007 4761 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.348138 4761 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.348219 4761 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.348336 4761 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.350451 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.351226 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29" (OuterVolumeSpecName: "kube-api-access-2lj29") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "kube-api-access-2lj29". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.357432 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" (UID: "f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.449752 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.449965 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450132 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-systemd-units\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450194 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-systemd-units\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450316 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450434 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-slash\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450479 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-config\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450523 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-bin\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450559 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-log-socket\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450638 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-env-overrides\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450744 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/005e52a3-41d1-438f-b439-f68b6585e596-ovn-node-metrics-cert\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450824 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-kubelet\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450905 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-netd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.450959 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-netns\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451043 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-script-lib\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451100 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-var-lib-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451143 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-systemd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451197 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-node-log\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451275 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-ovn\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451319 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzwpw\" (UniqueName: \"kubernetes.io/projected/005e52a3-41d1-438f-b439-f68b6585e596-kube-api-access-qzwpw\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451379 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-etc-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451439 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451562 4761 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451613 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lj29\" (UniqueName: \"kubernetes.io/projected/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-kube-api-access-2lj29\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451639 4761 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451692 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451779 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-slash\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.451913 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-netns\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452125 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-var-lib-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452180 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-systemd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452208 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-netd\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452240 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-kubelet\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452326 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-cni-bin\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452387 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-run-ovn\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452432 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-etc-openvswitch\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452138 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-node-log\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.452800 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-log-socket\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.453009 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-env-overrides\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.453195 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-config\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.453571 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/005e52a3-41d1-438f-b439-f68b6585e596-host-run-ovn-kubernetes\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.453808 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/005e52a3-41d1-438f-b439-f68b6585e596-ovnkube-script-lib\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.456919 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/005e52a3-41d1-438f-b439-f68b6585e596-ovn-node-metrics-cert\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.480065 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzwpw\" (UniqueName: \"kubernetes.io/projected/005e52a3-41d1-438f-b439-f68b6585e596-kube-api-access-qzwpw\") pod \"ovnkube-node-czxbc\" (UID: \"005e52a3-41d1-438f-b439-f68b6585e596\") " pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.554055 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.985896 4761 generic.go:334] "Generic (PLEG): container finished" podID="005e52a3-41d1-438f-b439-f68b6585e596" containerID="d6a8b957ee958d5c85b2820c3c5ce9ad97b902eb915f178569a24774607a3772" exitCode=0 Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.986022 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerDied","Data":"d6a8b957ee958d5c85b2820c3c5ce9ad97b902eb915f178569a24774607a3772"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.986085 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"3b3343872477ea042e31421fda7a1559277891f9f7223abf14195fd566672682"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.990005 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/2.log" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.994584 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-acl-logging/0.log" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995032 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qdk9h_f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/ovn-controller/0.log" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995358 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" exitCode=0 Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995385 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" exitCode=0 Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995393 4761 generic.go:334] "Generic (PLEG): container finished" podID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" exitCode=0 Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995420 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995433 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995453 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995468 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995480 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qdk9h" event={"ID":"f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8","Type":"ContainerDied","Data":"df94ff5676095a41933e5f37b2147d17ac04659d8ce4fff4beb6357d980be2fc"} Nov 25 23:06:30 crc kubenswrapper[4761]: I1125 23:06:30.995498 4761 scope.go:117] "RemoveContainer" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.018082 4761 scope.go:117] "RemoveContainer" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.031125 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qdk9h"] Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.036194 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qdk9h"] Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.048691 4761 scope.go:117] "RemoveContainer" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.076074 4761 scope.go:117] "RemoveContainer" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.101893 4761 scope.go:117] "RemoveContainer" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.117603 4761 scope.go:117] "RemoveContainer" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.134559 4761 scope.go:117] "RemoveContainer" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.156899 4761 scope.go:117] "RemoveContainer" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.187481 4761 scope.go:117] "RemoveContainer" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.227594 4761 scope.go:117] "RemoveContainer" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.228913 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": container with ID starting with 76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d not found: ID does not exist" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.228968 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d"} err="failed to get container status \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": rpc error: code = NotFound desc = could not find container \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": container with ID starting with 76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.229005 4761 scope.go:117] "RemoveContainer" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.229459 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": container with ID starting with e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3 not found: ID does not exist" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.229574 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3"} err="failed to get container status \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": rpc error: code = NotFound desc = could not find container \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": container with ID starting with e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.229627 4761 scope.go:117] "RemoveContainer" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.230915 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": container with ID starting with c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df not found: ID does not exist" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.230995 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df"} err="failed to get container status \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": rpc error: code = NotFound desc = could not find container \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": container with ID starting with c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.231029 4761 scope.go:117] "RemoveContainer" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.231516 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": container with ID starting with 6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20 not found: ID does not exist" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.231551 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20"} err="failed to get container status \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": rpc error: code = NotFound desc = could not find container \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": container with ID starting with 6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.231570 4761 scope.go:117] "RemoveContainer" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.231975 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": container with ID starting with 13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94 not found: ID does not exist" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232013 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94"} err="failed to get container status \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": rpc error: code = NotFound desc = could not find container \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": container with ID starting with 13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232040 4761 scope.go:117] "RemoveContainer" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.232400 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": container with ID starting with 507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1 not found: ID does not exist" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232436 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1"} err="failed to get container status \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": rpc error: code = NotFound desc = could not find container \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": container with ID starting with 507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232456 4761 scope.go:117] "RemoveContainer" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.232870 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": container with ID starting with b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c not found: ID does not exist" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232908 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c"} err="failed to get container status \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": rpc error: code = NotFound desc = could not find container \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": container with ID starting with b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.232929 4761 scope.go:117] "RemoveContainer" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.233284 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": container with ID starting with f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2 not found: ID does not exist" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.233318 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2"} err="failed to get container status \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": rpc error: code = NotFound desc = could not find container \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": container with ID starting with f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.233339 4761 scope.go:117] "RemoveContainer" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" Nov 25 23:06:31 crc kubenswrapper[4761]: E1125 23:06:31.233675 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": container with ID starting with 7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb not found: ID does not exist" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.233738 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb"} err="failed to get container status \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": rpc error: code = NotFound desc = could not find container \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": container with ID starting with 7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.233760 4761 scope.go:117] "RemoveContainer" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.234111 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d"} err="failed to get container status \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": rpc error: code = NotFound desc = could not find container \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": container with ID starting with 76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.234140 4761 scope.go:117] "RemoveContainer" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.234669 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3"} err="failed to get container status \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": rpc error: code = NotFound desc = could not find container \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": container with ID starting with e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.234721 4761 scope.go:117] "RemoveContainer" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235123 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df"} err="failed to get container status \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": rpc error: code = NotFound desc = could not find container \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": container with ID starting with c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235159 4761 scope.go:117] "RemoveContainer" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235507 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20"} err="failed to get container status \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": rpc error: code = NotFound desc = could not find container \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": container with ID starting with 6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235537 4761 scope.go:117] "RemoveContainer" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235944 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94"} err="failed to get container status \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": rpc error: code = NotFound desc = could not find container \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": container with ID starting with 13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.235969 4761 scope.go:117] "RemoveContainer" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.236395 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1"} err="failed to get container status \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": rpc error: code = NotFound desc = could not find container \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": container with ID starting with 507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.236419 4761 scope.go:117] "RemoveContainer" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.241221 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c"} err="failed to get container status \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": rpc error: code = NotFound desc = could not find container \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": container with ID starting with b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.241252 4761 scope.go:117] "RemoveContainer" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.242008 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2"} err="failed to get container status \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": rpc error: code = NotFound desc = could not find container \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": container with ID starting with f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.242070 4761 scope.go:117] "RemoveContainer" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.242585 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb"} err="failed to get container status \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": rpc error: code = NotFound desc = could not find container \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": container with ID starting with 7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.242621 4761 scope.go:117] "RemoveContainer" containerID="76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243125 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d"} err="failed to get container status \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": rpc error: code = NotFound desc = could not find container \"76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d\": container with ID starting with 76bce0306871695b6232411dd3da9b4d6ca88b36398c1074dfa850d007e2ff3d not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243162 4761 scope.go:117] "RemoveContainer" containerID="e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243450 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3"} err="failed to get container status \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": rpc error: code = NotFound desc = could not find container \"e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3\": container with ID starting with e5cfd3b1048a9ac2f6f1b0f4bc6d521bd8871283adc694c9bbfd6ef0c74c18e3 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243486 4761 scope.go:117] "RemoveContainer" containerID="c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243883 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df"} err="failed to get container status \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": rpc error: code = NotFound desc = could not find container \"c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df\": container with ID starting with c5bcd1fb723bc3bd04edc88bb51e1ec9c81c8c772ac5b1d3ac7688ebf22dd3df not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.243918 4761 scope.go:117] "RemoveContainer" containerID="6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.244148 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20"} err="failed to get container status \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": rpc error: code = NotFound desc = could not find container \"6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20\": container with ID starting with 6ef4743a4216f1db231d255c0afdee73eaa7f1faff2d4ae5aa171f055d1f2c20 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.244182 4761 scope.go:117] "RemoveContainer" containerID="13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.244495 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94"} err="failed to get container status \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": rpc error: code = NotFound desc = could not find container \"13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94\": container with ID starting with 13feaea4abd5e7a892fc65f66b04a291bb611fa18ba91fb9906668068454ae94 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.244837 4761 scope.go:117] "RemoveContainer" containerID="507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.245402 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1"} err="failed to get container status \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": rpc error: code = NotFound desc = could not find container \"507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1\": container with ID starting with 507f890640a575f747abba94f27c89327f530ac51afc4bb6d53b7d07e1e7f1e1 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.245437 4761 scope.go:117] "RemoveContainer" containerID="b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.246352 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c"} err="failed to get container status \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": rpc error: code = NotFound desc = could not find container \"b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c\": container with ID starting with b3463a131baa6b5564cdfb450373d12ffeb22c21a1e616de2974d190189c4d8c not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.246382 4761 scope.go:117] "RemoveContainer" containerID="f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.247377 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2"} err="failed to get container status \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": rpc error: code = NotFound desc = could not find container \"f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2\": container with ID starting with f1ab3cf582d0a8bc0713de9d0e245aa95493078e9ccb4c0cecffc6b5ea2bf0b2 not found: ID does not exist" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.247407 4761 scope.go:117] "RemoveContainer" containerID="7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb" Nov 25 23:06:31 crc kubenswrapper[4761]: I1125 23:06:31.247803 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb"} err="failed to get container status \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": rpc error: code = NotFound desc = could not find container \"7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb\": container with ID starting with 7a486b124315a05fe4af3f030a89e4b8bf62f89fe88c2f61da5170fbf8913bdb not found: ID does not exist" Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005362 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"7a6a85940ec4ef8e615ee9ef1ea12e56cc7c2ecf2a498aee9de47ddb7723c21c"} Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005409 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"c9248da4651d2dda5d1a9e578288a64a2a218414460075e3d479137f65c0fda7"} Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005425 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"53092b3949578540c7a1e8c2d409f361a468904c70f21d4241f1e7c600c3833a"} Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005437 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"00a662cfe0af566fce770f6d1cceb1a3b46a2cb86bdf7b879909bd744e66b11d"} Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005449 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"e4a6bd3e1537bd3be915185fa07ada355340d7f395d23315882603cd51872d6f"} Nov 25 23:06:32 crc kubenswrapper[4761]: I1125 23:06:32.005461 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"5994ceac93b68f92ce9869e75b2dcec6766a46e96047f9b617fe605a8c11e7b7"} Nov 25 23:06:33 crc kubenswrapper[4761]: I1125 23:06:33.023387 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8" path="/var/lib/kubelet/pods/f0ef866e-3c5b-4dfc-b5a2-61ee81e8bde8/volumes" Nov 25 23:06:35 crc kubenswrapper[4761]: I1125 23:06:35.042102 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"32e57317f152d55bad708234dc27b3dffa6ea269f43f95cc7a5004bb44ac9fc6"} Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.056341 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" event={"ID":"005e52a3-41d1-438f-b439-f68b6585e596","Type":"ContainerStarted","Data":"d3331be460a97fb9c183277e80527bc9b423d6a3536daa6270ca7223abb23006"} Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.056877 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.056962 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.057063 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.082265 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.082342 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:06:37 crc kubenswrapper[4761]: I1125 23:06:37.094318 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" podStartSLOduration=7.094303833 podStartE2EDuration="7.094303833s" podCreationTimestamp="2025-11-25 23:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:06:37.091905899 +0000 UTC m=+592.825251744" watchObservedRunningTime="2025-11-25 23:06:37.094303833 +0000 UTC m=+592.827649668" Nov 25 23:06:43 crc kubenswrapper[4761]: I1125 23:06:43.012103 4761 scope.go:117] "RemoveContainer" containerID="438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4" Nov 25 23:06:43 crc kubenswrapper[4761]: E1125 23:06:43.012911 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-fnqp9_openshift-multus(5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6)\"" pod="openshift-multus/multus-fnqp9" podUID="5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.366071 4761 scope.go:117] "RemoveContainer" containerID="1b65410ea38cfa2aa4dc22d04b55e31820051442721b8a5b265ff4481a81c1d4" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.394996 4761 scope.go:117] "RemoveContainer" containerID="04da4f347015efc0ec17c80a503fb19b3e8250314f034ddcbbefbcdb43794007" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.413616 4761 scope.go:117] "RemoveContainer" containerID="de3f01a5069fc82d87389504e83e6e18872ed6d9b139f7854cd1db9b53b19818" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.436478 4761 scope.go:117] "RemoveContainer" containerID="324804955ddf37f7b9d77ab0edaced44e478c83ba0048525d21b019598326385" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.449676 4761 scope.go:117] "RemoveContainer" containerID="4fe6aa41bdc86b3b13ca8f275614dc12ca6700fe7f4bb08d49e5ce85c69fa524" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.465656 4761 scope.go:117] "RemoveContainer" containerID="81cebfafc795ff5c559807ea30e125c130ce6a3c7de666a78ceddc2ef0bb5f51" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.498663 4761 scope.go:117] "RemoveContainer" containerID="d3ac805418d73f39bd407bc88b0ef28c7e96dc90cacb168af86acb2d86c25715" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.519582 4761 scope.go:117] "RemoveContainer" containerID="f75078a2a6e64770afc69f1dd0786a828a6d4128ad33dce9b35595a0b41c7d0b" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.544381 4761 scope.go:117] "RemoveContainer" containerID="0fe9f8ab217e8e8d4e5dc39f2606cdf9d452487de52232156b67ae3320627f8b" Nov 25 23:06:45 crc kubenswrapper[4761]: I1125 23:06:45.560777 4761 scope.go:117] "RemoveContainer" containerID="f77b20dcf8c626671498098d027e29237b8f067dbd3887146360ba33013dca80" Nov 25 23:06:51 crc kubenswrapper[4761]: I1125 23:06:51.128221 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:06:51 crc kubenswrapper[4761]: I1125 23:06:51.128913 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:06:51 crc kubenswrapper[4761]: I1125 23:06:51.128970 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:06:51 crc kubenswrapper[4761]: I1125 23:06:51.129597 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:06:51 crc kubenswrapper[4761]: I1125 23:06:51.129655 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c" gracePeriod=600 Nov 25 23:06:52 crc kubenswrapper[4761]: I1125 23:06:52.175770 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c" exitCode=0 Nov 25 23:06:52 crc kubenswrapper[4761]: I1125 23:06:52.175897 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c"} Nov 25 23:06:52 crc kubenswrapper[4761]: I1125 23:06:52.176501 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4"} Nov 25 23:06:52 crc kubenswrapper[4761]: I1125 23:06:52.176534 4761 scope.go:117] "RemoveContainer" containerID="ac4f843bd31b283c651b2a5f30fcb151b54820e5ccb50cfd617b963818126f3e" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.435582 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2"] Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.438777 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.443098 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.448028 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2"] Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.624557 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g78lx\" (UniqueName: \"kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.624849 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.624949 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.725993 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g78lx\" (UniqueName: \"kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.727630 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.727810 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.728483 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.728800 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.765813 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g78lx\" (UniqueName: \"kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: I1125 23:06:56.776649 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: E1125 23:06:56.813357 4761 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(19323140bf19b140325921eac4f58a1758ba204dd73879d99f7fa1850a4389ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 23:06:56 crc kubenswrapper[4761]: E1125 23:06:56.813483 4761 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(19323140bf19b140325921eac4f58a1758ba204dd73879d99f7fa1850a4389ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: E1125 23:06:56.813521 4761 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(19323140bf19b140325921eac4f58a1758ba204dd73879d99f7fa1850a4389ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:56 crc kubenswrapper[4761]: E1125 23:06:56.813597 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace(856fec2b-b44c-48c6-80a1-f71e87505db3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace(856fec2b-b44c-48c6-80a1-f71e87505db3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(19323140bf19b140325921eac4f58a1758ba204dd73879d99f7fa1850a4389ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" Nov 25 23:06:57 crc kubenswrapper[4761]: I1125 23:06:57.012519 4761 scope.go:117] "RemoveContainer" containerID="438e89d8bf5c802f129212428fb7f477cc3f0bde8d4970e1960c31810bba37a4" Nov 25 23:06:57 crc kubenswrapper[4761]: I1125 23:06:57.212458 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:57 crc kubenswrapper[4761]: I1125 23:06:57.213522 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:57 crc kubenswrapper[4761]: E1125 23:06:57.260769 4761 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(2bf4d9e73fe88c8160bd4cfe2219f6cf13a16d380e961c3e149938872046782f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 23:06:57 crc kubenswrapper[4761]: E1125 23:06:57.260849 4761 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(2bf4d9e73fe88c8160bd4cfe2219f6cf13a16d380e961c3e149938872046782f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:57 crc kubenswrapper[4761]: E1125 23:06:57.260877 4761 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(2bf4d9e73fe88c8160bd4cfe2219f6cf13a16d380e961c3e149938872046782f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:06:57 crc kubenswrapper[4761]: E1125 23:06:57.260939 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace(856fec2b-b44c-48c6-80a1-f71e87505db3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace(856fec2b-b44c-48c6-80a1-f71e87505db3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_openshift-marketplace_856fec2b-b44c-48c6-80a1-f71e87505db3_0(2bf4d9e73fe88c8160bd4cfe2219f6cf13a16d380e961c3e149938872046782f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" Nov 25 23:06:58 crc kubenswrapper[4761]: I1125 23:06:58.224166 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fnqp9_5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6/kube-multus/2.log" Nov 25 23:06:58 crc kubenswrapper[4761]: I1125 23:06:58.224236 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fnqp9" event={"ID":"5a2c9afd-91bb-4dd1-b4f2-6c139ffc19c6","Type":"ContainerStarted","Data":"421b28b3fba223ba9c23c40381a87ecb6dae4d025fbe92c675f3f98709e8fdd4"} Nov 25 23:07:00 crc kubenswrapper[4761]: I1125 23:07:00.589773 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-czxbc" Nov 25 23:07:10 crc kubenswrapper[4761]: I1125 23:07:10.011189 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:07:10 crc kubenswrapper[4761]: I1125 23:07:10.013841 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:07:10 crc kubenswrapper[4761]: I1125 23:07:10.282481 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2"] Nov 25 23:07:10 crc kubenswrapper[4761]: W1125 23:07:10.292906 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod856fec2b_b44c_48c6_80a1_f71e87505db3.slice/crio-f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f WatchSource:0}: Error finding container f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f: Status 404 returned error can't find the container with id f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f Nov 25 23:07:10 crc kubenswrapper[4761]: I1125 23:07:10.306656 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" event={"ID":"856fec2b-b44c-48c6-80a1-f71e87505db3","Type":"ContainerStarted","Data":"f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f"} Nov 25 23:07:11 crc kubenswrapper[4761]: I1125 23:07:11.317202 4761 generic.go:334] "Generic (PLEG): container finished" podID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerID="ee6e65836db7bcacd6ab978e704d2834b11b075f5bd9b139adff3af5042f4748" exitCode=0 Nov 25 23:07:11 crc kubenswrapper[4761]: I1125 23:07:11.317319 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" event={"ID":"856fec2b-b44c-48c6-80a1-f71e87505db3","Type":"ContainerDied","Data":"ee6e65836db7bcacd6ab978e704d2834b11b075f5bd9b139adff3af5042f4748"} Nov 25 23:07:11 crc kubenswrapper[4761]: I1125 23:07:11.322444 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 23:07:13 crc kubenswrapper[4761]: I1125 23:07:13.336816 4761 generic.go:334] "Generic (PLEG): container finished" podID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerID="7cc3308adef9fa9735e68159bd1edf28eff871ef24ae9d5c789e05716a1d9c89" exitCode=0 Nov 25 23:07:13 crc kubenswrapper[4761]: I1125 23:07:13.336883 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" event={"ID":"856fec2b-b44c-48c6-80a1-f71e87505db3","Type":"ContainerDied","Data":"7cc3308adef9fa9735e68159bd1edf28eff871ef24ae9d5c789e05716a1d9c89"} Nov 25 23:07:14 crc kubenswrapper[4761]: I1125 23:07:14.348415 4761 generic.go:334] "Generic (PLEG): container finished" podID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerID="c8044084d1c00c0063d674dd931754d9c6297d05277c6f79624eb8225b44c27d" exitCode=0 Nov 25 23:07:14 crc kubenswrapper[4761]: I1125 23:07:14.348482 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" event={"ID":"856fec2b-b44c-48c6-80a1-f71e87505db3","Type":"ContainerDied","Data":"c8044084d1c00c0063d674dd931754d9c6297d05277c6f79624eb8225b44c27d"} Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.707177 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.859529 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g78lx\" (UniqueName: \"kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx\") pod \"856fec2b-b44c-48c6-80a1-f71e87505db3\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.859766 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util\") pod \"856fec2b-b44c-48c6-80a1-f71e87505db3\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.859843 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle\") pod \"856fec2b-b44c-48c6-80a1-f71e87505db3\" (UID: \"856fec2b-b44c-48c6-80a1-f71e87505db3\") " Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.861408 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle" (OuterVolumeSpecName: "bundle") pod "856fec2b-b44c-48c6-80a1-f71e87505db3" (UID: "856fec2b-b44c-48c6-80a1-f71e87505db3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.866774 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx" (OuterVolumeSpecName: "kube-api-access-g78lx") pod "856fec2b-b44c-48c6-80a1-f71e87505db3" (UID: "856fec2b-b44c-48c6-80a1-f71e87505db3"). InnerVolumeSpecName "kube-api-access-g78lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.961448 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g78lx\" (UniqueName: \"kubernetes.io/projected/856fec2b-b44c-48c6-80a1-f71e87505db3-kube-api-access-g78lx\") on node \"crc\" DevicePath \"\"" Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.961498 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:07:15 crc kubenswrapper[4761]: I1125 23:07:15.982811 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util" (OuterVolumeSpecName: "util") pod "856fec2b-b44c-48c6-80a1-f71e87505db3" (UID: "856fec2b-b44c-48c6-80a1-f71e87505db3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:07:16 crc kubenswrapper[4761]: I1125 23:07:16.062609 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856fec2b-b44c-48c6-80a1-f71e87505db3-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:07:16 crc kubenswrapper[4761]: I1125 23:07:16.367420 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" event={"ID":"856fec2b-b44c-48c6-80a1-f71e87505db3","Type":"ContainerDied","Data":"f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f"} Nov 25 23:07:16 crc kubenswrapper[4761]: I1125 23:07:16.367822 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5e3479bec0aab53c573e4ee6522eba6f6791b45024b5bf952bf47e96ad9ab8f" Nov 25 23:07:16 crc kubenswrapper[4761]: I1125 23:07:16.367502 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.683601 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq"] Nov 25 23:07:29 crc kubenswrapper[4761]: E1125 23:07:29.684041 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="pull" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.684052 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="pull" Nov 25 23:07:29 crc kubenswrapper[4761]: E1125 23:07:29.684062 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="util" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.684067 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="util" Nov 25 23:07:29 crc kubenswrapper[4761]: E1125 23:07:29.684080 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="extract" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.684086 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="extract" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.684176 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="856fec2b-b44c-48c6-80a1-f71e87505db3" containerName="extract" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.684515 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.686152 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-k2qxk" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.686479 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.687074 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.687338 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.688909 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.699532 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq"] Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.846083 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-webhook-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.846142 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdt9m\" (UniqueName: \"kubernetes.io/projected/bca3f822-b2ce-401e-b219-52d29e761c49-kube-api-access-fdt9m\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.846176 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-apiservice-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.904658 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7"] Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.905712 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.907119 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.907726 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-x4c9s" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.908277 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.919004 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7"] Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.947016 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-webhook-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.947326 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdt9m\" (UniqueName: \"kubernetes.io/projected/bca3f822-b2ce-401e-b219-52d29e761c49-kube-api-access-fdt9m\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.947414 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-apiservice-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.952565 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-apiservice-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.955726 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bca3f822-b2ce-401e-b219-52d29e761c49-webhook-cert\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:29 crc kubenswrapper[4761]: I1125 23:07:29.969199 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdt9m\" (UniqueName: \"kubernetes.io/projected/bca3f822-b2ce-401e-b219-52d29e761c49-kube-api-access-fdt9m\") pod \"metallb-operator-controller-manager-6d87477c79-vr8fq\" (UID: \"bca3f822-b2ce-401e-b219-52d29e761c49\") " pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.003948 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.049026 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbjk6\" (UniqueName: \"kubernetes.io/projected/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-kube-api-access-lbjk6\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.049074 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-apiservice-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.049106 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-webhook-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.150357 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbjk6\" (UniqueName: \"kubernetes.io/projected/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-kube-api-access-lbjk6\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.150398 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-apiservice-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.150434 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-webhook-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.155272 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-webhook-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.156335 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-apiservice-cert\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.170227 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbjk6\" (UniqueName: \"kubernetes.io/projected/98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23-kube-api-access-lbjk6\") pod \"metallb-operator-webhook-server-55745dcbf5-4xft7\" (UID: \"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23\") " pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.220756 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.243657 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq"] Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.400622 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7"] Nov 25 23:07:30 crc kubenswrapper[4761]: W1125 23:07:30.408943 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98ea4dd9_aefd_42d7_9a3a_c21aa5dbef23.slice/crio-859273cbc46f6a89bd2ddfd75c9d1e609ced5bd8a96064012e8f8b5dcea1285b WatchSource:0}: Error finding container 859273cbc46f6a89bd2ddfd75c9d1e609ced5bd8a96064012e8f8b5dcea1285b: Status 404 returned error can't find the container with id 859273cbc46f6a89bd2ddfd75c9d1e609ced5bd8a96064012e8f8b5dcea1285b Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.581805 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerStarted","Data":"f854c7963c7f647d166bd66dbef0428893f8c35ac7e831810f680a5bec3301e0"} Nov 25 23:07:30 crc kubenswrapper[4761]: I1125 23:07:30.583109 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" event={"ID":"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23","Type":"ContainerStarted","Data":"859273cbc46f6a89bd2ddfd75c9d1e609ced5bd8a96064012e8f8b5dcea1285b"} Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.611220 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" event={"ID":"98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23","Type":"ContainerStarted","Data":"f2903084c0c71723393ce174f4b6514cd2cac090a9c37bb3bb542c04d4b30be5"} Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.611658 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.613274 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerStarted","Data":"61d3817e1eb40c038638d44c35a68380ac5d3f980e9129cbc735b3a11dbc55a3"} Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.613440 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.643137 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" podStartSLOduration=2.081381346 podStartE2EDuration="6.643123234s" podCreationTimestamp="2025-11-25 23:07:29 +0000 UTC" firstStartedPulling="2025-11-25 23:07:30.410896533 +0000 UTC m=+646.144242358" lastFinishedPulling="2025-11-25 23:07:34.972638411 +0000 UTC m=+650.705984246" observedRunningTime="2025-11-25 23:07:35.636830136 +0000 UTC m=+651.370175981" watchObservedRunningTime="2025-11-25 23:07:35.643123234 +0000 UTC m=+651.376469069" Nov 25 23:07:35 crc kubenswrapper[4761]: I1125 23:07:35.664595 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" podStartSLOduration=1.9687127960000002 podStartE2EDuration="6.664571807s" podCreationTimestamp="2025-11-25 23:07:29 +0000 UTC" firstStartedPulling="2025-11-25 23:07:30.258156051 +0000 UTC m=+645.991501886" lastFinishedPulling="2025-11-25 23:07:34.954015062 +0000 UTC m=+650.687360897" observedRunningTime="2025-11-25 23:07:35.662165062 +0000 UTC m=+651.395510927" watchObservedRunningTime="2025-11-25 23:07:35.664571807 +0000 UTC m=+651.397917642" Nov 25 23:07:50 crc kubenswrapper[4761]: I1125 23:07:50.236097 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-55745dcbf5-4xft7" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.008088 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.872077 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-l78lp"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.874929 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.877292 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-c9dlg" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.877334 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.878752 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.883283 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-shxv2"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.884655 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.886663 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.896800 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-shxv2"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955339 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22mlm\" (UniqueName: \"kubernetes.io/projected/8220142c-79e5-43f5-9580-e88813af96cd-kube-api-access-22mlm\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955405 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-sockets\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955426 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955443 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-startup\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955463 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955482 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-conf\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955777 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-reloader\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955856 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics-certs\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.955901 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ckzw\" (UniqueName: \"kubernetes.io/projected/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-kube-api-access-4ckzw\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.966299 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-sbf9j"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.967609 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sbf9j" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.969647 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.970292 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.971512 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2cpjb" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.981807 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.987894 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-ftvpd"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.989188 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-ftvpd"] Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.989293 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:10 crc kubenswrapper[4761]: I1125 23:08:10.991177 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057511 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22mlm\" (UniqueName: \"kubernetes.io/projected/8220142c-79e5-43f5-9580-e88813af96cd-kube-api-access-22mlm\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057579 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-sockets\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057602 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057626 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057644 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-startup\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057662 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057684 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7lkz\" (UniqueName: \"kubernetes.io/projected/a6d03a2c-b764-4a2d-a55d-91cb3a608412-kube-api-access-n7lkz\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057714 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-conf\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057738 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-metrics-certs\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057761 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzlbg\" (UniqueName: \"kubernetes.io/projected/a441d0d3-7366-499e-8096-2877837d1e58-kube-api-access-tzlbg\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057781 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-reloader\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057804 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-cert\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057823 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics-certs\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057840 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ckzw\" (UniqueName: \"kubernetes.io/projected/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-kube-api-access-4ckzw\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057862 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a441d0d3-7366-499e-8096-2877837d1e58-metallb-excludel2\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.057881 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.057993 4761 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.058052 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert podName:8220142c-79e5-43f5-9580-e88813af96cd nodeName:}" failed. No retries permitted until 2025-11-25 23:08:11.55802742 +0000 UTC m=+687.291373255 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert") pod "frr-k8s-webhook-server-6998585d5-shxv2" (UID: "8220142c-79e5-43f5-9580-e88813af96cd") : secret "frr-k8s-webhook-server-cert" not found Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.058141 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-sockets\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.058262 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-reloader\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.058411 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.058650 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-conf\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.058915 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-frr-startup\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.063417 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-metrics-certs\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.073945 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ckzw\" (UniqueName: \"kubernetes.io/projected/8973d4eb-45ec-49e9-a487-3963ae1b2fb9-kube-api-access-4ckzw\") pod \"frr-k8s-l78lp\" (UID: \"8973d4eb-45ec-49e9-a487-3963ae1b2fb9\") " pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.083114 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22mlm\" (UniqueName: \"kubernetes.io/projected/8220142c-79e5-43f5-9580-e88813af96cd-kube-api-access-22mlm\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159325 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7lkz\" (UniqueName: \"kubernetes.io/projected/a6d03a2c-b764-4a2d-a55d-91cb3a608412-kube-api-access-n7lkz\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159591 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-metrics-certs\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159681 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzlbg\" (UniqueName: \"kubernetes.io/projected/a441d0d3-7366-499e-8096-2877837d1e58-kube-api-access-tzlbg\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159792 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-cert\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159879 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.159949 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a441d0d3-7366-499e-8096-2877837d1e58-metallb-excludel2\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.160058 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.159959 4761 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.160270 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs podName:a441d0d3-7366-499e-8096-2877837d1e58 nodeName:}" failed. No retries permitted until 2025-11-25 23:08:11.660254781 +0000 UTC m=+687.393600616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs") pod "speaker-sbf9j" (UID: "a441d0d3-7366-499e-8096-2877837d1e58") : secret "speaker-certs-secret" not found Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.160133 4761 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.160394 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist podName:a441d0d3-7366-499e-8096-2877837d1e58 nodeName:}" failed. No retries permitted until 2025-11-25 23:08:11.660387234 +0000 UTC m=+687.393733059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist") pod "speaker-sbf9j" (UID: "a441d0d3-7366-499e-8096-2877837d1e58") : secret "metallb-memberlist" not found Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.160514 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a441d0d3-7366-499e-8096-2877837d1e58-metallb-excludel2\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.163297 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.165214 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-metrics-certs\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.176000 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6d03a2c-b764-4a2d-a55d-91cb3a608412-cert\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.178633 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7lkz\" (UniqueName: \"kubernetes.io/projected/a6d03a2c-b764-4a2d-a55d-91cb3a608412-kube-api-access-n7lkz\") pod \"controller-6c7b4b5f48-ftvpd\" (UID: \"a6d03a2c-b764-4a2d-a55d-91cb3a608412\") " pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.185908 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzlbg\" (UniqueName: \"kubernetes.io/projected/a441d0d3-7366-499e-8096-2877837d1e58-kube-api-access-tzlbg\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.208993 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.301669 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.531791 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-ftvpd"] Nov 25 23:08:11 crc kubenswrapper[4761]: W1125 23:08:11.547180 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6d03a2c_b764_4a2d_a55d_91cb3a608412.slice/crio-59cb830a40b1d4baf8a77325559c47a8c5fc63ab63ff415a1204b8e7f9cddf7a WatchSource:0}: Error finding container 59cb830a40b1d4baf8a77325559c47a8c5fc63ab63ff415a1204b8e7f9cddf7a: Status 404 returned error can't find the container with id 59cb830a40b1d4baf8a77325559c47a8c5fc63ab63ff415a1204b8e7f9cddf7a Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.568973 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.575908 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8220142c-79e5-43f5-9580-e88813af96cd-cert\") pod \"frr-k8s-webhook-server-6998585d5-shxv2\" (UID: \"8220142c-79e5-43f5-9580-e88813af96cd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.670290 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.670344 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.670446 4761 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 23:08:11 crc kubenswrapper[4761]: E1125 23:08:11.670497 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist podName:a441d0d3-7366-499e-8096-2877837d1e58 nodeName:}" failed. No retries permitted until 2025-11-25 23:08:12.670481773 +0000 UTC m=+688.403827608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist") pod "speaker-sbf9j" (UID: "a441d0d3-7366-499e-8096-2877837d1e58") : secret "metallb-memberlist" not found Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.673763 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-metrics-certs\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.825030 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.853049 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"7d0e2fc7d97299197940f5b3e28af899ff7c9df74a709f3994ba2b61badd970b"} Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.854926 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-ftvpd" event={"ID":"a6d03a2c-b764-4a2d-a55d-91cb3a608412","Type":"ContainerStarted","Data":"0a356016ebf8f879bea60c5a18a0b0524596f1597b7d8004aef777041fe8e29f"} Nov 25 23:08:11 crc kubenswrapper[4761]: I1125 23:08:11.854964 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-ftvpd" event={"ID":"a6d03a2c-b764-4a2d-a55d-91cb3a608412","Type":"ContainerStarted","Data":"59cb830a40b1d4baf8a77325559c47a8c5fc63ab63ff415a1204b8e7f9cddf7a"} Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.064992 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-shxv2"] Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.683971 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.706559 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a441d0d3-7366-499e-8096-2877837d1e58-memberlist\") pod \"speaker-sbf9j\" (UID: \"a441d0d3-7366-499e-8096-2877837d1e58\") " pod="metallb-system/speaker-sbf9j" Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.779765 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sbf9j" Nov 25 23:08:12 crc kubenswrapper[4761]: W1125 23:08:12.805633 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda441d0d3_7366_499e_8096_2877837d1e58.slice/crio-6259e24b95220922de3586e9b3df0d1d16e5dd8e71c719682f1f25d0432e2f73 WatchSource:0}: Error finding container 6259e24b95220922de3586e9b3df0d1d16e5dd8e71c719682f1f25d0432e2f73: Status 404 returned error can't find the container with id 6259e24b95220922de3586e9b3df0d1d16e5dd8e71c719682f1f25d0432e2f73 Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.867022 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" event={"ID":"8220142c-79e5-43f5-9580-e88813af96cd","Type":"ContainerStarted","Data":"78acd0c974f1fe4330ccf3ef8783547867a7d611fcace4e299947d13de3f14bc"} Nov 25 23:08:12 crc kubenswrapper[4761]: I1125 23:08:12.870930 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sbf9j" event={"ID":"a441d0d3-7366-499e-8096-2877837d1e58","Type":"ContainerStarted","Data":"6259e24b95220922de3586e9b3df0d1d16e5dd8e71c719682f1f25d0432e2f73"} Nov 25 23:08:13 crc kubenswrapper[4761]: I1125 23:08:13.887792 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sbf9j" event={"ID":"a441d0d3-7366-499e-8096-2877837d1e58","Type":"ContainerStarted","Data":"3aa8c271a84f705b8bd096aa3c684c05417b8e5153c166cde5ee11b1277770fe"} Nov 25 23:08:16 crc kubenswrapper[4761]: I1125 23:08:16.921566 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-ftvpd" event={"ID":"a6d03a2c-b764-4a2d-a55d-91cb3a608412","Type":"ContainerStarted","Data":"8eb7567cd2596aec3fb4c743de20a9663873df780edd372f636edd52d9495498"} Nov 25 23:08:16 crc kubenswrapper[4761]: I1125 23:08:16.922164 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:16 crc kubenswrapper[4761]: I1125 23:08:16.925104 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sbf9j" event={"ID":"a441d0d3-7366-499e-8096-2877837d1e58","Type":"ContainerStarted","Data":"59e750887a3c307fe301b2f7dd1b5e8ea2acc8ec223974cd98db506ed6f182d5"} Nov 25 23:08:16 crc kubenswrapper[4761]: I1125 23:08:16.925686 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-sbf9j" Nov 25 23:08:16 crc kubenswrapper[4761]: I1125 23:08:16.936771 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-ftvpd" podStartSLOduration=2.655653534 podStartE2EDuration="6.936753293s" podCreationTimestamp="2025-11-25 23:08:10 +0000 UTC" firstStartedPulling="2025-11-25 23:08:11.668064889 +0000 UTC m=+687.401410724" lastFinishedPulling="2025-11-25 23:08:15.949164648 +0000 UTC m=+691.682510483" observedRunningTime="2025-11-25 23:08:16.934931555 +0000 UTC m=+692.668277400" watchObservedRunningTime="2025-11-25 23:08:16.936753293 +0000 UTC m=+692.670099138" Nov 25 23:08:19 crc kubenswrapper[4761]: I1125 23:08:19.945275 4761 generic.go:334] "Generic (PLEG): container finished" podID="8973d4eb-45ec-49e9-a487-3963ae1b2fb9" containerID="47144adb225a26dd89a40df8f15878273e080beca7e69b480027c65fcd6001fd" exitCode=0 Nov 25 23:08:19 crc kubenswrapper[4761]: I1125 23:08:19.945337 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerDied","Data":"47144adb225a26dd89a40df8f15878273e080beca7e69b480027c65fcd6001fd"} Nov 25 23:08:19 crc kubenswrapper[4761]: I1125 23:08:19.950834 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" event={"ID":"8220142c-79e5-43f5-9580-e88813af96cd","Type":"ContainerStarted","Data":"6924b098b6d7f440c09733c651a4c067416badfa855383750e469c261e022194"} Nov 25 23:08:19 crc kubenswrapper[4761]: I1125 23:08:19.951147 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:19 crc kubenswrapper[4761]: I1125 23:08:19.991041 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-sbf9j" podStartSLOduration=7.158326963 podStartE2EDuration="9.991015764s" podCreationTimestamp="2025-11-25 23:08:10 +0000 UTC" firstStartedPulling="2025-11-25 23:08:13.114096973 +0000 UTC m=+688.847442828" lastFinishedPulling="2025-11-25 23:08:15.946785794 +0000 UTC m=+691.680131629" observedRunningTime="2025-11-25 23:08:16.950808778 +0000 UTC m=+692.684154613" watchObservedRunningTime="2025-11-25 23:08:19.991015764 +0000 UTC m=+695.724361629" Nov 25 23:08:20 crc kubenswrapper[4761]: I1125 23:08:20.011015 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" podStartSLOduration=2.757282229 podStartE2EDuration="10.010986938s" podCreationTimestamp="2025-11-25 23:08:10 +0000 UTC" firstStartedPulling="2025-11-25 23:08:12.073074069 +0000 UTC m=+687.806419904" lastFinishedPulling="2025-11-25 23:08:19.326778748 +0000 UTC m=+695.060124613" observedRunningTime="2025-11-25 23:08:20.010750712 +0000 UTC m=+695.744096577" watchObservedRunningTime="2025-11-25 23:08:20.010986938 +0000 UTC m=+695.744332813" Nov 25 23:08:20 crc kubenswrapper[4761]: I1125 23:08:20.963358 4761 generic.go:334] "Generic (PLEG): container finished" podID="8973d4eb-45ec-49e9-a487-3963ae1b2fb9" containerID="9faed87c95eb4635dd088469c3a2529c1f25249761ff9ada89727ebfd462f179" exitCode=0 Nov 25 23:08:20 crc kubenswrapper[4761]: I1125 23:08:20.963469 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerDied","Data":"9faed87c95eb4635dd088469c3a2529c1f25249761ff9ada89727ebfd462f179"} Nov 25 23:08:21 crc kubenswrapper[4761]: I1125 23:08:21.307907 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-ftvpd" Nov 25 23:08:21 crc kubenswrapper[4761]: I1125 23:08:21.975765 4761 generic.go:334] "Generic (PLEG): container finished" podID="8973d4eb-45ec-49e9-a487-3963ae1b2fb9" containerID="b29f0bcd6020c2098429424cfacc5c57dfa0431a1dcc76a0b09e61f37c46c759" exitCode=0 Nov 25 23:08:21 crc kubenswrapper[4761]: I1125 23:08:21.975854 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerDied","Data":"b29f0bcd6020c2098429424cfacc5c57dfa0431a1dcc76a0b09e61f37c46c759"} Nov 25 23:08:22 crc kubenswrapper[4761]: I1125 23:08:22.991662 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"c89ecc45f6d0149f2f2fc55cb177ef3e9527608da1555be2d90fdb9f412f9161"} Nov 25 23:08:22 crc kubenswrapper[4761]: I1125 23:08:22.992275 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"1a3c860415091f816febf7e759630752a5e05b21fd26d60aed6e709196c2f507"} Nov 25 23:08:22 crc kubenswrapper[4761]: I1125 23:08:22.992302 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"61ce3e6b228056e3b5147a3da82bafa16f67cf94b30553bc199c99494b172dc2"} Nov 25 23:08:22 crc kubenswrapper[4761]: I1125 23:08:22.992320 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"394aae57b529e5d0d09df9bf184ab8259e7727f89daf3998487b8e4feb0fd420"} Nov 25 23:08:22 crc kubenswrapper[4761]: I1125 23:08:22.992338 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"de1e9fcbb44bc2fb82c0af5f8f2534528c0ecf48ab3ec22c2c0b96b5d0019b61"} Nov 25 23:08:24 crc kubenswrapper[4761]: I1125 23:08:24.007660 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l78lp" event={"ID":"8973d4eb-45ec-49e9-a487-3963ae1b2fb9","Type":"ContainerStarted","Data":"33c088e69f59fbbcc4290263bae53f96c111b61d59682e7ae492df5ca0a1b8fe"} Nov 25 23:08:24 crc kubenswrapper[4761]: I1125 23:08:24.008164 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:24 crc kubenswrapper[4761]: I1125 23:08:24.047155 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-l78lp" podStartSLOduration=6.000807536 podStartE2EDuration="14.047132032s" podCreationTimestamp="2025-11-25 23:08:10 +0000 UTC" firstStartedPulling="2025-11-25 23:08:11.303831507 +0000 UTC m=+687.037177342" lastFinishedPulling="2025-11-25 23:08:19.350155973 +0000 UTC m=+695.083501838" observedRunningTime="2025-11-25 23:08:24.042628172 +0000 UTC m=+699.775974097" watchObservedRunningTime="2025-11-25 23:08:24.047132032 +0000 UTC m=+699.780477897" Nov 25 23:08:26 crc kubenswrapper[4761]: I1125 23:08:26.209372 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:26 crc kubenswrapper[4761]: I1125 23:08:26.293317 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:31 crc kubenswrapper[4761]: I1125 23:08:31.828871 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-shxv2" Nov 25 23:08:32 crc kubenswrapper[4761]: I1125 23:08:32.784837 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-sbf9j" Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.776938 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.778763 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.788225 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.788644 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-5qvmg" Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.810854 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.816807 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:38 crc kubenswrapper[4761]: I1125 23:08:38.911129 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bggl8\" (UniqueName: \"kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8\") pod \"mariadb-operator-index-d76ms\" (UID: \"56adeae3-19ea-4eec-b4b0-695971443bb5\") " pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:39 crc kubenswrapper[4761]: I1125 23:08:39.012018 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bggl8\" (UniqueName: \"kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8\") pod \"mariadb-operator-index-d76ms\" (UID: \"56adeae3-19ea-4eec-b4b0-695971443bb5\") " pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:39 crc kubenswrapper[4761]: I1125 23:08:39.033253 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bggl8\" (UniqueName: \"kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8\") pod \"mariadb-operator-index-d76ms\" (UID: \"56adeae3-19ea-4eec-b4b0-695971443bb5\") " pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:39 crc kubenswrapper[4761]: I1125 23:08:39.109014 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:39 crc kubenswrapper[4761]: I1125 23:08:39.398664 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:40 crc kubenswrapper[4761]: I1125 23:08:40.129182 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d76ms" event={"ID":"56adeae3-19ea-4eec-b4b0-695971443bb5","Type":"ContainerStarted","Data":"62c5b657caa7b1a51178b0969837d6ed0143aa51a47c33d38f37bfd022b23fce"} Nov 25 23:08:41 crc kubenswrapper[4761]: I1125 23:08:41.140369 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d76ms" event={"ID":"56adeae3-19ea-4eec-b4b0-695971443bb5","Type":"ContainerStarted","Data":"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468"} Nov 25 23:08:41 crc kubenswrapper[4761]: I1125 23:08:41.167026 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-d76ms" podStartSLOduration=2.205550257 podStartE2EDuration="3.166999199s" podCreationTimestamp="2025-11-25 23:08:38 +0000 UTC" firstStartedPulling="2025-11-25 23:08:39.41210346 +0000 UTC m=+715.145449335" lastFinishedPulling="2025-11-25 23:08:40.373552432 +0000 UTC m=+716.106898277" observedRunningTime="2025-11-25 23:08:41.165053456 +0000 UTC m=+716.898399331" watchObservedRunningTime="2025-11-25 23:08:41.166999199 +0000 UTC m=+716.900345064" Nov 25 23:08:41 crc kubenswrapper[4761]: I1125 23:08:41.218099 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-l78lp" Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.142847 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.772198 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-tq4fc"] Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.773008 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.802294 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-tq4fc"] Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.872949 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbpt\" (UniqueName: \"kubernetes.io/projected/a74e9524-d069-42d8-9e98-ec3eed11ec22-kube-api-access-4kbpt\") pod \"mariadb-operator-index-tq4fc\" (UID: \"a74e9524-d069-42d8-9e98-ec3eed11ec22\") " pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.974710 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbpt\" (UniqueName: \"kubernetes.io/projected/a74e9524-d069-42d8-9e98-ec3eed11ec22-kube-api-access-4kbpt\") pod \"mariadb-operator-index-tq4fc\" (UID: \"a74e9524-d069-42d8-9e98-ec3eed11ec22\") " pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:42 crc kubenswrapper[4761]: I1125 23:08:42.994138 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbpt\" (UniqueName: \"kubernetes.io/projected/a74e9524-d069-42d8-9e98-ec3eed11ec22-kube-api-access-4kbpt\") pod \"mariadb-operator-index-tq4fc\" (UID: \"a74e9524-d069-42d8-9e98-ec3eed11ec22\") " pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.095952 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.153029 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-d76ms" podUID="56adeae3-19ea-4eec-b4b0-695971443bb5" containerName="registry-server" containerID="cri-o://b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468" gracePeriod=2 Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.507064 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.575510 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-tq4fc"] Nov 25 23:08:43 crc kubenswrapper[4761]: W1125 23:08:43.583115 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda74e9524_d069_42d8_9e98_ec3eed11ec22.slice/crio-82afd95f274f3932a0d1656a498b149c6e3e8e8abb57d4fb74de92dcf9e73fe4 WatchSource:0}: Error finding container 82afd95f274f3932a0d1656a498b149c6e3e8e8abb57d4fb74de92dcf9e73fe4: Status 404 returned error can't find the container with id 82afd95f274f3932a0d1656a498b149c6e3e8e8abb57d4fb74de92dcf9e73fe4 Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.583390 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bggl8\" (UniqueName: \"kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8\") pod \"56adeae3-19ea-4eec-b4b0-695971443bb5\" (UID: \"56adeae3-19ea-4eec-b4b0-695971443bb5\") " Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.590320 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8" (OuterVolumeSpecName: "kube-api-access-bggl8") pod "56adeae3-19ea-4eec-b4b0-695971443bb5" (UID: "56adeae3-19ea-4eec-b4b0-695971443bb5"). InnerVolumeSpecName "kube-api-access-bggl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:08:43 crc kubenswrapper[4761]: I1125 23:08:43.684894 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bggl8\" (UniqueName: \"kubernetes.io/projected/56adeae3-19ea-4eec-b4b0-695971443bb5-kube-api-access-bggl8\") on node \"crc\" DevicePath \"\"" Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.160983 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-tq4fc" event={"ID":"a74e9524-d069-42d8-9e98-ec3eed11ec22","Type":"ContainerStarted","Data":"82afd95f274f3932a0d1656a498b149c6e3e8e8abb57d4fb74de92dcf9e73fe4"} Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.163910 4761 generic.go:334] "Generic (PLEG): container finished" podID="56adeae3-19ea-4eec-b4b0-695971443bb5" containerID="b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468" exitCode=0 Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.163982 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d76ms" Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.164023 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d76ms" event={"ID":"56adeae3-19ea-4eec-b4b0-695971443bb5","Type":"ContainerDied","Data":"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468"} Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.164533 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d76ms" event={"ID":"56adeae3-19ea-4eec-b4b0-695971443bb5","Type":"ContainerDied","Data":"62c5b657caa7b1a51178b0969837d6ed0143aa51a47c33d38f37bfd022b23fce"} Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.164585 4761 scope.go:117] "RemoveContainer" containerID="b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468" Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.189925 4761 scope.go:117] "RemoveContainer" containerID="b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468" Nov 25 23:08:44 crc kubenswrapper[4761]: E1125 23:08:44.190428 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468\": container with ID starting with b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468 not found: ID does not exist" containerID="b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468" Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.190503 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468"} err="failed to get container status \"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468\": rpc error: code = NotFound desc = could not find container \"b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468\": container with ID starting with b40a174c09a74585657f0ea0123c0335324d9df601f59e6079a686d085209468 not found: ID does not exist" Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.210354 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:44 crc kubenswrapper[4761]: I1125 23:08:44.216530 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-d76ms"] Nov 25 23:08:45 crc kubenswrapper[4761]: I1125 23:08:45.021970 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56adeae3-19ea-4eec-b4b0-695971443bb5" path="/var/lib/kubelet/pods/56adeae3-19ea-4eec-b4b0-695971443bb5/volumes" Nov 25 23:08:45 crc kubenswrapper[4761]: I1125 23:08:45.173188 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-tq4fc" event={"ID":"a74e9524-d069-42d8-9e98-ec3eed11ec22","Type":"ContainerStarted","Data":"b77fec1c0630fab1488afffb9047e6891215d26e5a88f0bc9ea13b6746006d58"} Nov 25 23:08:45 crc kubenswrapper[4761]: I1125 23:08:45.201778 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-tq4fc" podStartSLOduration=2.20632452 podStartE2EDuration="3.201752225s" podCreationTimestamp="2025-11-25 23:08:42 +0000 UTC" firstStartedPulling="2025-11-25 23:08:43.588252825 +0000 UTC m=+719.321598670" lastFinishedPulling="2025-11-25 23:08:44.5836805 +0000 UTC m=+720.317026375" observedRunningTime="2025-11-25 23:08:45.201640682 +0000 UTC m=+720.934986507" watchObservedRunningTime="2025-11-25 23:08:45.201752225 +0000 UTC m=+720.935098090" Nov 25 23:08:51 crc kubenswrapper[4761]: I1125 23:08:51.128289 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:08:51 crc kubenswrapper[4761]: I1125 23:08:51.128882 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:08:53 crc kubenswrapper[4761]: I1125 23:08:53.097890 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:53 crc kubenswrapper[4761]: I1125 23:08:53.098349 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:53 crc kubenswrapper[4761]: I1125 23:08:53.145637 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:53 crc kubenswrapper[4761]: I1125 23:08:53.282719 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-tq4fc" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.266004 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts"] Nov 25 23:08:59 crc kubenswrapper[4761]: E1125 23:08:59.266641 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56adeae3-19ea-4eec-b4b0-695971443bb5" containerName="registry-server" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.266661 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="56adeae3-19ea-4eec-b4b0-695971443bb5" containerName="registry-server" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.266878 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="56adeae3-19ea-4eec-b4b0-695971443bb5" containerName="registry-server" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.268141 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.270826 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.279127 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts"] Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.317197 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.317277 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.317358 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r78t7\" (UniqueName: \"kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.418144 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.418557 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r78t7\" (UniqueName: \"kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.418614 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.418786 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.419215 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.453683 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r78t7\" (UniqueName: \"kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:08:59 crc kubenswrapper[4761]: I1125 23:08:59.587474 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:09:00 crc kubenswrapper[4761]: I1125 23:09:00.107280 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts"] Nov 25 23:09:00 crc kubenswrapper[4761]: W1125 23:09:00.117814 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15aaf702_bd98_4dbd_b5e7_296a7546cd5a.slice/crio-7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540 WatchSource:0}: Error finding container 7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540: Status 404 returned error can't find the container with id 7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540 Nov 25 23:09:00 crc kubenswrapper[4761]: I1125 23:09:00.294373 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" event={"ID":"15aaf702-bd98-4dbd-b5e7-296a7546cd5a","Type":"ContainerStarted","Data":"7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540"} Nov 25 23:09:01 crc kubenswrapper[4761]: I1125 23:09:01.304604 4761 generic.go:334] "Generic (PLEG): container finished" podID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerID="6f87c378e6a2b4f1fe82305f0045cca80ea94cee07d8557277d331159a5b7a80" exitCode=0 Nov 25 23:09:01 crc kubenswrapper[4761]: I1125 23:09:01.304668 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" event={"ID":"15aaf702-bd98-4dbd-b5e7-296a7546cd5a","Type":"ContainerDied","Data":"6f87c378e6a2b4f1fe82305f0045cca80ea94cee07d8557277d331159a5b7a80"} Nov 25 23:09:03 crc kubenswrapper[4761]: I1125 23:09:03.324208 4761 generic.go:334] "Generic (PLEG): container finished" podID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerID="801de5817344b15bd18224e6231af8a28e20570991bc0f9f8448f537f1d75ce3" exitCode=0 Nov 25 23:09:03 crc kubenswrapper[4761]: I1125 23:09:03.324299 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" event={"ID":"15aaf702-bd98-4dbd-b5e7-296a7546cd5a","Type":"ContainerDied","Data":"801de5817344b15bd18224e6231af8a28e20570991bc0f9f8448f537f1d75ce3"} Nov 25 23:09:04 crc kubenswrapper[4761]: I1125 23:09:04.335949 4761 generic.go:334] "Generic (PLEG): container finished" podID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerID="05a708f1b5102f0fe795e2bc63f70b682243bb0234c7313a8792950621a47d34" exitCode=0 Nov 25 23:09:04 crc kubenswrapper[4761]: I1125 23:09:04.336009 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" event={"ID":"15aaf702-bd98-4dbd-b5e7-296a7546cd5a","Type":"ContainerDied","Data":"05a708f1b5102f0fe795e2bc63f70b682243bb0234c7313a8792950621a47d34"} Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.644608 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.715879 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r78t7\" (UniqueName: \"kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7\") pod \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.716091 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle\") pod \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.716155 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util\") pod \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\" (UID: \"15aaf702-bd98-4dbd-b5e7-296a7546cd5a\") " Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.717756 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle" (OuterVolumeSpecName: "bundle") pod "15aaf702-bd98-4dbd-b5e7-296a7546cd5a" (UID: "15aaf702-bd98-4dbd-b5e7-296a7546cd5a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.732768 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7" (OuterVolumeSpecName: "kube-api-access-r78t7") pod "15aaf702-bd98-4dbd-b5e7-296a7546cd5a" (UID: "15aaf702-bd98-4dbd-b5e7-296a7546cd5a"). InnerVolumeSpecName "kube-api-access-r78t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.818636 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:05 crc kubenswrapper[4761]: I1125 23:09:05.818685 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r78t7\" (UniqueName: \"kubernetes.io/projected/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-kube-api-access-r78t7\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:06 crc kubenswrapper[4761]: I1125 23:09:06.043857 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util" (OuterVolumeSpecName: "util") pod "15aaf702-bd98-4dbd-b5e7-296a7546cd5a" (UID: "15aaf702-bd98-4dbd-b5e7-296a7546cd5a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:09:06 crc kubenswrapper[4761]: I1125 23:09:06.123463 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15aaf702-bd98-4dbd-b5e7-296a7546cd5a-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:06 crc kubenswrapper[4761]: I1125 23:09:06.354676 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" event={"ID":"15aaf702-bd98-4dbd-b5e7-296a7546cd5a","Type":"ContainerDied","Data":"7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540"} Nov 25 23:09:06 crc kubenswrapper[4761]: I1125 23:09:06.354786 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a16765a271c5f8427ec3b70037c5c04b82addb3b4a257b2d7859daed45b5540" Nov 25 23:09:06 crc kubenswrapper[4761]: I1125 23:09:06.354802 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts" Nov 25 23:09:10 crc kubenswrapper[4761]: I1125 23:09:10.799395 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 23:09:10 crc kubenswrapper[4761]: I1125 23:09:10.800195 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerName="controller-manager" containerID="cri-o://354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f" gracePeriod=30 Nov 25 23:09:10 crc kubenswrapper[4761]: I1125 23:09:10.902903 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 23:09:10 crc kubenswrapper[4761]: I1125 23:09:10.903160 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" podUID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" containerName="route-controller-manager" containerID="cri-o://b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1" gracePeriod=30 Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.163045 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.236842 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296774 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca\") pod \"d7a16660-730c-4eb8-bfca-6ee890409ba0\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296841 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config\") pod \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296882 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config\") pod \"d7a16660-730c-4eb8-bfca-6ee890409ba0\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296911 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nclxk\" (UniqueName: \"kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk\") pod \"d7a16660-730c-4eb8-bfca-6ee890409ba0\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296935 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert\") pod \"d7a16660-730c-4eb8-bfca-6ee890409ba0\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296950 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca\") pod \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.296965 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert\") pod \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.297013 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles\") pod \"d7a16660-730c-4eb8-bfca-6ee890409ba0\" (UID: \"d7a16660-730c-4eb8-bfca-6ee890409ba0\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.297386 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca" (OuterVolumeSpecName: "client-ca") pod "d7a16660-730c-4eb8-bfca-6ee890409ba0" (UID: "d7a16660-730c-4eb8-bfca-6ee890409ba0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.298220 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d7a16660-730c-4eb8-bfca-6ee890409ba0" (UID: "d7a16660-730c-4eb8-bfca-6ee890409ba0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.298269 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca" (OuterVolumeSpecName: "client-ca") pod "b5471e18-fe28-4269-bbd1-7e93dbd45b22" (UID: "b5471e18-fe28-4269-bbd1-7e93dbd45b22"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.298658 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config" (OuterVolumeSpecName: "config") pod "b5471e18-fe28-4269-bbd1-7e93dbd45b22" (UID: "b5471e18-fe28-4269-bbd1-7e93dbd45b22"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.299589 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config" (OuterVolumeSpecName: "config") pod "d7a16660-730c-4eb8-bfca-6ee890409ba0" (UID: "d7a16660-730c-4eb8-bfca-6ee890409ba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.302711 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk" (OuterVolumeSpecName: "kube-api-access-nclxk") pod "d7a16660-730c-4eb8-bfca-6ee890409ba0" (UID: "d7a16660-730c-4eb8-bfca-6ee890409ba0"). InnerVolumeSpecName "kube-api-access-nclxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.303813 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7a16660-730c-4eb8-bfca-6ee890409ba0" (UID: "d7a16660-730c-4eb8-bfca-6ee890409ba0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.303903 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b5471e18-fe28-4269-bbd1-7e93dbd45b22" (UID: "b5471e18-fe28-4269-bbd1-7e93dbd45b22"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.390890 4761 generic.go:334] "Generic (PLEG): container finished" podID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerID="354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f" exitCode=0 Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.390949 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" event={"ID":"d7a16660-730c-4eb8-bfca-6ee890409ba0","Type":"ContainerDied","Data":"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f"} Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.390987 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.391220 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9p7tc" event={"ID":"d7a16660-730c-4eb8-bfca-6ee890409ba0","Type":"ContainerDied","Data":"8dbf497d525f2e3d09575a2602425f639f493613bb23c1fedd1d4af93ed25408"} Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.391253 4761 scope.go:117] "RemoveContainer" containerID="354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.392560 4761 generic.go:334] "Generic (PLEG): container finished" podID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" containerID="b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1" exitCode=0 Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.392594 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" event={"ID":"b5471e18-fe28-4269-bbd1-7e93dbd45b22","Type":"ContainerDied","Data":"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1"} Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.392614 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" event={"ID":"b5471e18-fe28-4269-bbd1-7e93dbd45b22","Type":"ContainerDied","Data":"9b3ddcd8dc9e61131de61b8d7e1bb6e9ba077491ac280ef856830c9012eadbb3"} Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.392785 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.397795 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29rbd\" (UniqueName: \"kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd\") pod \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\" (UID: \"b5471e18-fe28-4269-bbd1-7e93dbd45b22\") " Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.397979 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-config\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.397997 4761 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-config\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398006 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nclxk\" (UniqueName: \"kubernetes.io/projected/d7a16660-730c-4eb8-bfca-6ee890409ba0-kube-api-access-nclxk\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398018 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7a16660-730c-4eb8-bfca-6ee890409ba0-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398026 4761 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5471e18-fe28-4269-bbd1-7e93dbd45b22-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398035 4761 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5471e18-fe28-4269-bbd1-7e93dbd45b22-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398043 4761 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.398052 4761 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7a16660-730c-4eb8-bfca-6ee890409ba0-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.403125 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd" (OuterVolumeSpecName: "kube-api-access-29rbd") pod "b5471e18-fe28-4269-bbd1-7e93dbd45b22" (UID: "b5471e18-fe28-4269-bbd1-7e93dbd45b22"). InnerVolumeSpecName "kube-api-access-29rbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.409762 4761 scope.go:117] "RemoveContainer" containerID="354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f" Nov 25 23:09:11 crc kubenswrapper[4761]: E1125 23:09:11.410234 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f\": container with ID starting with 354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f not found: ID does not exist" containerID="354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.410310 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f"} err="failed to get container status \"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f\": rpc error: code = NotFound desc = could not find container \"354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f\": container with ID starting with 354ba95694091f68ceb49e69211e3270df7fe156f151428e7c1671b4feb4312f not found: ID does not exist" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.410343 4761 scope.go:117] "RemoveContainer" containerID="b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.432025 4761 scope.go:117] "RemoveContainer" containerID="b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1" Nov 25 23:09:11 crc kubenswrapper[4761]: E1125 23:09:11.432508 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1\": container with ID starting with b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1 not found: ID does not exist" containerID="b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.432552 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1"} err="failed to get container status \"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1\": rpc error: code = NotFound desc = could not find container \"b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1\": container with ID starting with b0563f1ffb9688449f3588acb56138e4e91786f9e03e4e879244221476df80f1 not found: ID does not exist" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.434369 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.441467 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9p7tc"] Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.499005 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29rbd\" (UniqueName: \"kubernetes.io/projected/b5471e18-fe28-4269-bbd1-7e93dbd45b22-kube-api-access-29rbd\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.726985 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 23:09:11 crc kubenswrapper[4761]: I1125 23:09:11.730698 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l4nsv"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169585 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42"] Nov 25 23:09:12 crc kubenswrapper[4761]: E1125 23:09:12.169791 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" containerName="route-controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169803 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" containerName="route-controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: E1125 23:09:12.169814 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="pull" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169819 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="pull" Nov 25 23:09:12 crc kubenswrapper[4761]: E1125 23:09:12.169826 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="extract" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169832 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="extract" Nov 25 23:09:12 crc kubenswrapper[4761]: E1125 23:09:12.169840 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="util" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169846 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="util" Nov 25 23:09:12 crc kubenswrapper[4761]: E1125 23:09:12.169858 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerName="controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169863 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerName="controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169949 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="15aaf702-bd98-4dbd-b5e7-296a7546cd5a" containerName="extract" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169969 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" containerName="route-controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.169979 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" containerName="controller-manager" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.170339 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.174160 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.174483 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.221339 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.308072 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-apiservice-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.308157 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xbc\" (UniqueName: \"kubernetes.io/projected/0717f7d1-d5b3-408e-878e-10c62251448a-kube-api-access-97xbc\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.308303 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-webhook-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.320064 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75578895b6-qgv7l"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.320664 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.323228 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.323297 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.323318 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.323973 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.324512 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.324717 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.328723 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.329331 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.333624 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.335102 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.335797 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.335907 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.336033 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.336139 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.340133 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.340967 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.368342 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75578895b6-qgv7l"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.409655 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xbc\" (UniqueName: \"kubernetes.io/projected/0717f7d1-d5b3-408e-878e-10c62251448a-kube-api-access-97xbc\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.410009 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-webhook-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.413200 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-apiservice-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.419473 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-webhook-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.419514 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0717f7d1-d5b3-408e-878e-10c62251448a-apiservice-cert\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.434508 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xbc\" (UniqueName: \"kubernetes.io/projected/0717f7d1-d5b3-408e-878e-10c62251448a-kube-api-access-97xbc\") pod \"mariadb-operator-controller-manager-6bbbcc9f68-dcp42\" (UID: \"0717f7d1-d5b3-408e-878e-10c62251448a\") " pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.489529 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514467 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-config\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514786 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfa934a-c400-4d00-b135-a632fa52b5f6-serving-cert\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514814 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-config\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514870 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck68b\" (UniqueName: \"kubernetes.io/projected/ddfa934a-c400-4d00-b135-a632fa52b5f6-kube-api-access-ck68b\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514891 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-proxy-ca-bundles\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514942 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrpt7\" (UniqueName: \"kubernetes.io/projected/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-kube-api-access-rrpt7\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514967 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-client-ca\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.514989 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-client-ca\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.515012 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-serving-cert\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616347 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-client-ca\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616405 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-client-ca\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616432 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-serving-cert\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616470 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-config\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616492 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfa934a-c400-4d00-b135-a632fa52b5f6-serving-cert\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616513 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-config\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616528 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck68b\" (UniqueName: \"kubernetes.io/projected/ddfa934a-c400-4d00-b135-a632fa52b5f6-kube-api-access-ck68b\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616546 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-proxy-ca-bundles\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.616563 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrpt7\" (UniqueName: \"kubernetes.io/projected/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-kube-api-access-rrpt7\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.617690 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-client-ca\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.618305 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-client-ca\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.619788 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-config\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.621012 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-config\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.622451 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddfa934a-c400-4d00-b135-a632fa52b5f6-proxy-ca-bundles\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.624920 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-serving-cert\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.625249 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfa934a-c400-4d00-b135-a632fa52b5f6-serving-cert\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.639936 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrpt7\" (UniqueName: \"kubernetes.io/projected/59a7d61d-64d3-4e29-b6b6-0b114b85b69c-kube-api-access-rrpt7\") pod \"route-controller-manager-5b7456b476-b6nmn\" (UID: \"59a7d61d-64d3-4e29-b6b6-0b114b85b69c\") " pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.641263 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck68b\" (UniqueName: \"kubernetes.io/projected/ddfa934a-c400-4d00-b135-a632fa52b5f6-kube-api-access-ck68b\") pod \"controller-manager-75578895b6-qgv7l\" (UID: \"ddfa934a-c400-4d00-b135-a632fa52b5f6\") " pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.647136 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.710785 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42"] Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.879641 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn"] Nov 25 23:09:12 crc kubenswrapper[4761]: W1125 23:09:12.885524 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a7d61d_64d3_4e29_b6b6_0b114b85b69c.slice/crio-d9385d1a4a7b49e5d440e2be7f19c5baeb14980913fcc0c5057ec2af4db41c96 WatchSource:0}: Error finding container d9385d1a4a7b49e5d440e2be7f19c5baeb14980913fcc0c5057ec2af4db41c96: Status 404 returned error can't find the container with id d9385d1a4a7b49e5d440e2be7f19c5baeb14980913fcc0c5057ec2af4db41c96 Nov 25 23:09:12 crc kubenswrapper[4761]: I1125 23:09:12.932706 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.019369 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5471e18-fe28-4269-bbd1-7e93dbd45b22" path="/var/lib/kubelet/pods/b5471e18-fe28-4269-bbd1-7e93dbd45b22/volumes" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.020066 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a16660-730c-4eb8-bfca-6ee890409ba0" path="/var/lib/kubelet/pods/d7a16660-730c-4eb8-bfca-6ee890409ba0/volumes" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.204909 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75578895b6-qgv7l"] Nov 25 23:09:13 crc kubenswrapper[4761]: W1125 23:09:13.207208 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddfa934a_c400_4d00_b135_a632fa52b5f6.slice/crio-333b1ba26442d6b2cc72612d9af9152b3c3c069bff92f84b6f1e3ba6ad6594af WatchSource:0}: Error finding container 333b1ba26442d6b2cc72612d9af9152b3c3c069bff92f84b6f1e3ba6ad6594af: Status 404 returned error can't find the container with id 333b1ba26442d6b2cc72612d9af9152b3c3c069bff92f84b6f1e3ba6ad6594af Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.424581 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerStarted","Data":"51f40464b201cc9197ec337eba8af5182abc71af0ede3f45c573c82fd6a4f609"} Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.429616 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" event={"ID":"ddfa934a-c400-4d00-b135-a632fa52b5f6","Type":"ContainerStarted","Data":"fd68813e2c82c8c7650c79dc36295558f18051231fae096403e40453bd04bd65"} Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.429668 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" event={"ID":"ddfa934a-c400-4d00-b135-a632fa52b5f6","Type":"ContainerStarted","Data":"333b1ba26442d6b2cc72612d9af9152b3c3c069bff92f84b6f1e3ba6ad6594af"} Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.431681 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.431781 4761 patch_prober.go:28] interesting pod/controller-manager-75578895b6-qgv7l container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.431816 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" podUID="ddfa934a-c400-4d00-b135-a632fa52b5f6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.439052 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" event={"ID":"59a7d61d-64d3-4e29-b6b6-0b114b85b69c","Type":"ContainerStarted","Data":"899c2416b189a2820099f9acf384eee35ae4b463307ef2d5d28dd315ef06873a"} Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.439091 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" event={"ID":"59a7d61d-64d3-4e29-b6b6-0b114b85b69c","Type":"ContainerStarted","Data":"d9385d1a4a7b49e5d440e2be7f19c5baeb14980913fcc0c5057ec2af4db41c96"} Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.439479 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.448180 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.453609 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" podStartSLOduration=1.453584602 podStartE2EDuration="1.453584602s" podCreationTimestamp="2025-11-25 23:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:09:13.448407374 +0000 UTC m=+749.181753219" watchObservedRunningTime="2025-11-25 23:09:13.453584602 +0000 UTC m=+749.186930437" Nov 25 23:09:13 crc kubenswrapper[4761]: I1125 23:09:13.474869 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b7456b476-b6nmn" podStartSLOduration=1.474851957 podStartE2EDuration="1.474851957s" podCreationTimestamp="2025-11-25 23:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:09:13.472534026 +0000 UTC m=+749.205879881" watchObservedRunningTime="2025-11-25 23:09:13.474851957 +0000 UTC m=+749.208197792" Nov 25 23:09:14 crc kubenswrapper[4761]: I1125 23:09:14.452185 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75578895b6-qgv7l" Nov 25 23:09:17 crc kubenswrapper[4761]: I1125 23:09:17.467780 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerStarted","Data":"31d225e37138e67eda4e68aee35ffa55d0cc8952747907f352a150fd953f45ff"} Nov 25 23:09:17 crc kubenswrapper[4761]: I1125 23:09:17.468736 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:17 crc kubenswrapper[4761]: I1125 23:09:17.490442 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podStartSLOduration=1.78689307 podStartE2EDuration="5.490403124s" podCreationTimestamp="2025-11-25 23:09:12 +0000 UTC" firstStartedPulling="2025-11-25 23:09:12.724147986 +0000 UTC m=+748.457493821" lastFinishedPulling="2025-11-25 23:09:16.42765804 +0000 UTC m=+752.161003875" observedRunningTime="2025-11-25 23:09:17.488465293 +0000 UTC m=+753.221811228" watchObservedRunningTime="2025-11-25 23:09:17.490403124 +0000 UTC m=+753.223749009" Nov 25 23:09:19 crc kubenswrapper[4761]: I1125 23:09:19.391550 4761 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 23:09:21 crc kubenswrapper[4761]: I1125 23:09:21.128379 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:09:21 crc kubenswrapper[4761]: I1125 23:09:21.128492 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:09:22 crc kubenswrapper[4761]: I1125 23:09:22.500486 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:09:24 crc kubenswrapper[4761]: I1125 23:09:24.975909 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-4qknl"] Nov 25 23:09:24 crc kubenswrapper[4761]: I1125 23:09:24.977062 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:24 crc kubenswrapper[4761]: I1125 23:09:24.979226 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-fjnq8" Nov 25 23:09:24 crc kubenswrapper[4761]: I1125 23:09:24.983909 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4qknl"] Nov 25 23:09:25 crc kubenswrapper[4761]: I1125 23:09:25.002634 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fwbb\" (UniqueName: \"kubernetes.io/projected/7fdf0c47-5b58-4cac-8fe2-df4f2e58d797-kube-api-access-2fwbb\") pod \"infra-operator-index-4qknl\" (UID: \"7fdf0c47-5b58-4cac-8fe2-df4f2e58d797\") " pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:25 crc kubenswrapper[4761]: I1125 23:09:25.103635 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fwbb\" (UniqueName: \"kubernetes.io/projected/7fdf0c47-5b58-4cac-8fe2-df4f2e58d797-kube-api-access-2fwbb\") pod \"infra-operator-index-4qknl\" (UID: \"7fdf0c47-5b58-4cac-8fe2-df4f2e58d797\") " pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:25 crc kubenswrapper[4761]: I1125 23:09:25.132402 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fwbb\" (UniqueName: \"kubernetes.io/projected/7fdf0c47-5b58-4cac-8fe2-df4f2e58d797-kube-api-access-2fwbb\") pod \"infra-operator-index-4qknl\" (UID: \"7fdf0c47-5b58-4cac-8fe2-df4f2e58d797\") " pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:25 crc kubenswrapper[4761]: I1125 23:09:25.294067 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:25 crc kubenswrapper[4761]: I1125 23:09:25.766832 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4qknl"] Nov 25 23:09:26 crc kubenswrapper[4761]: I1125 23:09:26.519251 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4qknl" event={"ID":"7fdf0c47-5b58-4cac-8fe2-df4f2e58d797","Type":"ContainerStarted","Data":"308fc2d8cc8f1c5f5599461f1fb6c992ff3f1d9f03808dc09466c5ee0d2d9140"} Nov 25 23:09:27 crc kubenswrapper[4761]: I1125 23:09:27.530321 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4qknl" event={"ID":"7fdf0c47-5b58-4cac-8fe2-df4f2e58d797","Type":"ContainerStarted","Data":"cfc060e1c846fae55c5e1b829fbed0fd4c185aedc3f8823e05f1ec6db3e7f640"} Nov 25 23:09:27 crc kubenswrapper[4761]: I1125 23:09:27.557988 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-4qknl" podStartSLOduration=2.49730013 podStartE2EDuration="3.557962088s" podCreationTimestamp="2025-11-25 23:09:24 +0000 UTC" firstStartedPulling="2025-11-25 23:09:25.782245106 +0000 UTC m=+761.515590972" lastFinishedPulling="2025-11-25 23:09:26.842907095 +0000 UTC m=+762.576252930" observedRunningTime="2025-11-25 23:09:27.552436472 +0000 UTC m=+763.285782367" watchObservedRunningTime="2025-11-25 23:09:27.557962088 +0000 UTC m=+763.291307963" Nov 25 23:09:35 crc kubenswrapper[4761]: I1125 23:09:35.295176 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:35 crc kubenswrapper[4761]: I1125 23:09:35.295802 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:35 crc kubenswrapper[4761]: I1125 23:09:35.339494 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:35 crc kubenswrapper[4761]: I1125 23:09:35.625559 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-4qknl" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.245637 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw"] Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.248178 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.252172 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.261412 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw"] Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.350797 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.350893 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.350991 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vptgh\" (UniqueName: \"kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.452853 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.452958 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.453095 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vptgh\" (UniqueName: \"kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.454512 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.455069 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.528636 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vptgh\" (UniqueName: \"kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:42 crc kubenswrapper[4761]: I1125 23:09:42.575082 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:43 crc kubenswrapper[4761]: I1125 23:09:43.068438 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw"] Nov 25 23:09:43 crc kubenswrapper[4761]: W1125 23:09:43.075472 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod205db7be_4e2b_4785_85f4_b5a76f277139.slice/crio-94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83 WatchSource:0}: Error finding container 94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83: Status 404 returned error can't find the container with id 94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83 Nov 25 23:09:43 crc kubenswrapper[4761]: I1125 23:09:43.657644 4761 generic.go:334] "Generic (PLEG): container finished" podID="205db7be-4e2b-4785-85f4-b5a76f277139" containerID="21a2be8937aead50c53d2e328c08c18ad94921394d77ae7591205ae1661bc985" exitCode=0 Nov 25 23:09:43 crc kubenswrapper[4761]: I1125 23:09:43.657848 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerDied","Data":"21a2be8937aead50c53d2e328c08c18ad94921394d77ae7591205ae1661bc985"} Nov 25 23:09:43 crc kubenswrapper[4761]: I1125 23:09:43.658087 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerStarted","Data":"94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83"} Nov 25 23:09:44 crc kubenswrapper[4761]: I1125 23:09:44.670276 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerStarted","Data":"46d29aefb0b74d30916ae3fa8006784d7d080f9465dc8c0bf463851db8fd3ec6"} Nov 25 23:09:45 crc kubenswrapper[4761]: I1125 23:09:45.681072 4761 generic.go:334] "Generic (PLEG): container finished" podID="205db7be-4e2b-4785-85f4-b5a76f277139" containerID="46d29aefb0b74d30916ae3fa8006784d7d080f9465dc8c0bf463851db8fd3ec6" exitCode=0 Nov 25 23:09:45 crc kubenswrapper[4761]: I1125 23:09:45.681923 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerDied","Data":"46d29aefb0b74d30916ae3fa8006784d7d080f9465dc8c0bf463851db8fd3ec6"} Nov 25 23:09:46 crc kubenswrapper[4761]: I1125 23:09:46.692242 4761 generic.go:334] "Generic (PLEG): container finished" podID="205db7be-4e2b-4785-85f4-b5a76f277139" containerID="371611305a19230295e5ea82f56947175da4eee06a06f6b2cf495059f3f61ff2" exitCode=0 Nov 25 23:09:46 crc kubenswrapper[4761]: I1125 23:09:46.692379 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerDied","Data":"371611305a19230295e5ea82f56947175da4eee06a06f6b2cf495059f3f61ff2"} Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.061338 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.139983 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vptgh\" (UniqueName: \"kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh\") pod \"205db7be-4e2b-4785-85f4-b5a76f277139\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.140059 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle\") pod \"205db7be-4e2b-4785-85f4-b5a76f277139\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.140131 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util\") pod \"205db7be-4e2b-4785-85f4-b5a76f277139\" (UID: \"205db7be-4e2b-4785-85f4-b5a76f277139\") " Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.141539 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle" (OuterVolumeSpecName: "bundle") pod "205db7be-4e2b-4785-85f4-b5a76f277139" (UID: "205db7be-4e2b-4785-85f4-b5a76f277139"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.147575 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh" (OuterVolumeSpecName: "kube-api-access-vptgh") pod "205db7be-4e2b-4785-85f4-b5a76f277139" (UID: "205db7be-4e2b-4785-85f4-b5a76f277139"). InnerVolumeSpecName "kube-api-access-vptgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.162185 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util" (OuterVolumeSpecName: "util") pod "205db7be-4e2b-4785-85f4-b5a76f277139" (UID: "205db7be-4e2b-4785-85f4-b5a76f277139"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.241996 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vptgh\" (UniqueName: \"kubernetes.io/projected/205db7be-4e2b-4785-85f4-b5a76f277139-kube-api-access-vptgh\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.242051 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.242071 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205db7be-4e2b-4785-85f4-b5a76f277139-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.707856 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" event={"ID":"205db7be-4e2b-4785-85f4-b5a76f277139","Type":"ContainerDied","Data":"94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83"} Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.707896 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94fdfbbe0e963e256db62992590c449b09c24d164cfae85fdb43c75401440b83" Nov 25 23:09:48 crc kubenswrapper[4761]: I1125 23:09:48.707936 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw" Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.127987 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.128406 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.128470 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.129398 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.129526 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4" gracePeriod=600 Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.753051 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4" exitCode=0 Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.753186 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4"} Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.753683 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8"} Nov 25 23:09:51 crc kubenswrapper[4761]: I1125 23:09:51.753756 4761 scope.go:117] "RemoveContainer" containerID="8877ae377805c4b6f75be4b99b110edede6bf97cf6f4b795460ad0613d6e555c" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.353199 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7"] Nov 25 23:09:53 crc kubenswrapper[4761]: E1125 23:09:53.354491 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="util" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.354551 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="util" Nov 25 23:09:53 crc kubenswrapper[4761]: E1125 23:09:53.354575 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="extract" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.354584 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="extract" Nov 25 23:09:53 crc kubenswrapper[4761]: E1125 23:09:53.354616 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="pull" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.354623 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="pull" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.355025 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="205db7be-4e2b-4785-85f4-b5a76f277139" containerName="extract" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.356005 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.369868 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kjg7n" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.369868 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.386984 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7"] Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.410909 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9ptq\" (UniqueName: \"kubernetes.io/projected/2e34c701-6bbc-4000-a90a-aeca2333b69b-kube-api-access-m9ptq\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.410968 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-apiservice-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.412912 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-webhook-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.514038 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9ptq\" (UniqueName: \"kubernetes.io/projected/2e34c701-6bbc-4000-a90a-aeca2333b69b-kube-api-access-m9ptq\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.514092 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-apiservice-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.514129 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-webhook-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.521418 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-webhook-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.529735 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9ptq\" (UniqueName: \"kubernetes.io/projected/2e34c701-6bbc-4000-a90a-aeca2333b69b-kube-api-access-m9ptq\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.533581 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e34c701-6bbc-4000-a90a-aeca2333b69b-apiservice-cert\") pod \"infra-operator-controller-manager-5c5dd6c96-z8jr7\" (UID: \"2e34c701-6bbc-4000-a90a-aeca2333b69b\") " pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:53 crc kubenswrapper[4761]: I1125 23:09:53.674232 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:54 crc kubenswrapper[4761]: I1125 23:09:54.164157 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7"] Nov 25 23:09:54 crc kubenswrapper[4761]: I1125 23:09:54.783358 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerStarted","Data":"539679abc3c213e78215b61d4293b7c5a0590bf1a82204702c419cf8640ef057"} Nov 25 23:09:56 crc kubenswrapper[4761]: I1125 23:09:56.803064 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerStarted","Data":"9567de82eec7b39af5f0974fe4a4370e520ded7e082979e868d5438cfa39451a"} Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.233753 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.241382 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.246934 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.247253 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.247718 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-ch5fs" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.247827 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.247996 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.248805 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.276044 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.278974 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285401 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5cf936d6-7508-40de-a1ed-537da8087430-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285476 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-kolla-config\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285529 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgbbf\" (UniqueName: \"kubernetes.io/projected/5cf936d6-7508-40de-a1ed-537da8087430-kube-api-access-pgbbf\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285620 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285737 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.285813 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-config-data-default\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.286963 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.289048 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.294550 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.306676 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386610 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t46p\" (UniqueName: \"kubernetes.io/projected/3c5a3c85-d95b-4252-8feb-155b5b742101-kube-api-access-4t46p\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386667 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5cf936d6-7508-40de-a1ed-537da8087430-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386688 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386730 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-kolla-config\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386749 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgbbf\" (UniqueName: \"kubernetes.io/projected/5cf936d6-7508-40de-a1ed-537da8087430-kube-api-access-pgbbf\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386769 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-kolla-config\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386788 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386822 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386847 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-default\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386878 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386903 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-config-data-default\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.386935 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.387520 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5cf936d6-7508-40de-a1ed-537da8087430-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.387527 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.388033 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-kolla-config\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.388076 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-config-data-default\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.389501 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cf936d6-7508-40de-a1ed-537da8087430-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.410185 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.412036 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgbbf\" (UniqueName: \"kubernetes.io/projected/5cf936d6-7508-40de-a1ed-537da8087430-kube-api-access-pgbbf\") pod \"openstack-galera-0\" (UID: \"5cf936d6-7508-40de-a1ed-537da8087430\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488172 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-default\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488214 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488238 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488259 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488326 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-kolla-config\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488407 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-config-data-default\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488433 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488464 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t46p\" (UniqueName: \"kubernetes.io/projected/3c5a3c85-d95b-4252-8feb-155b5b742101-kube-api-access-4t46p\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488494 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488513 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d770ea4c-d11c-4526-af47-7507131606c0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488533 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-kolla-config\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488554 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl8lk\" (UniqueName: \"kubernetes.io/projected/d770ea4c-d11c-4526-af47-7507131606c0-kube-api-access-xl8lk\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.488773 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.489452 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.489465 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-config-data-default\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.489839 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-kolla-config\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.490332 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c5a3c85-d95b-4252-8feb-155b5b742101-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.516103 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.534570 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t46p\" (UniqueName: \"kubernetes.io/projected/3c5a3c85-d95b-4252-8feb-155b5b742101-kube-api-access-4t46p\") pod \"openstack-galera-1\" (UID: \"3c5a3c85-d95b-4252-8feb-155b5b742101\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.579827 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590235 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-kolla-config\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590283 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-config-data-default\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590333 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d770ea4c-d11c-4526-af47-7507131606c0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590356 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl8lk\" (UniqueName: \"kubernetes.io/projected/d770ea4c-d11c-4526-af47-7507131606c0-kube-api-access-xl8lk\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590387 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590403 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590534 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.590947 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-kolla-config\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.591109 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-config-data-default\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.591434 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d770ea4c-d11c-4526-af47-7507131606c0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.592102 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d770ea4c-d11c-4526-af47-7507131606c0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.610036 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.615476 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl8lk\" (UniqueName: \"kubernetes.io/projected/d770ea4c-d11c-4526-af47-7507131606c0-kube-api-access-xl8lk\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.616240 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-2\" (UID: \"d770ea4c-d11c-4526-af47-7507131606c0\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.626004 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.816218 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerStarted","Data":"940abb6622a9ea13d87134d7018fafa543f0b5ba4f65c79189b85d8e8c1255a4"} Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.816663 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.842906 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podStartSLOduration=1.652576338 podStartE2EDuration="5.842888162s" podCreationTimestamp="2025-11-25 23:09:53 +0000 UTC" firstStartedPulling="2025-11-25 23:09:54.177442942 +0000 UTC m=+789.910788777" lastFinishedPulling="2025-11-25 23:09:58.367754746 +0000 UTC m=+794.101100601" observedRunningTime="2025-11-25 23:09:58.841120165 +0000 UTC m=+794.574466000" watchObservedRunningTime="2025-11-25 23:09:58.842888162 +0000 UTC m=+794.576233997" Nov 25 23:09:58 crc kubenswrapper[4761]: I1125 23:09:58.971977 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 23:09:58 crc kubenswrapper[4761]: W1125 23:09:58.980192 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c5a3c85_d95b_4252_8feb_155b5b742101.slice/crio-a54fa417e744402de6afc7569c6b6c527cce09e4e27e9fabb8176e6cac6479d3 WatchSource:0}: Error finding container a54fa417e744402de6afc7569c6b6c527cce09e4e27e9fabb8176e6cac6479d3: Status 404 returned error can't find the container with id a54fa417e744402de6afc7569c6b6c527cce09e4e27e9fabb8176e6cac6479d3 Nov 25 23:09:59 crc kubenswrapper[4761]: I1125 23:09:59.063599 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 23:09:59 crc kubenswrapper[4761]: I1125 23:09:59.131459 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 23:09:59 crc kubenswrapper[4761]: I1125 23:09:59.825199 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d770ea4c-d11c-4526-af47-7507131606c0","Type":"ContainerStarted","Data":"e6b38caebc8559b3b543821c33d3b84e869622f169e7f7f295027459c7b1f0b8"} Nov 25 23:09:59 crc kubenswrapper[4761]: I1125 23:09:59.826198 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5cf936d6-7508-40de-a1ed-537da8087430","Type":"ContainerStarted","Data":"2457801cfb1d41750727edb3951b828867d55850e2f2e13bb47917ec850ca0a6"} Nov 25 23:09:59 crc kubenswrapper[4761]: I1125 23:09:59.827640 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3c5a3c85-d95b-4252-8feb-155b5b742101","Type":"ContainerStarted","Data":"a54fa417e744402de6afc7569c6b6c527cce09e4e27e9fabb8176e6cac6479d3"} Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.582542 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.586339 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.590040 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.645086 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.645134 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqw9n\" (UniqueName: \"kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.645252 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.746439 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.746488 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqw9n\" (UniqueName: \"kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.746563 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.747113 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.747124 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.774960 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqw9n\" (UniqueName: \"kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n\") pod \"community-operators-zpjtv\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:01 crc kubenswrapper[4761]: I1125 23:10:01.916129 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:03 crc kubenswrapper[4761]: I1125 23:10:03.680475 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.579499 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.580955 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.596490 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.690467 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmkxl\" (UniqueName: \"kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.690562 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.690586 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.792218 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.792285 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.792338 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmkxl\" (UniqueName: \"kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.794205 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.794981 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.824382 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmkxl\" (UniqueName: \"kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl\") pod \"redhat-marketplace-8nmjp\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:04 crc kubenswrapper[4761]: I1125 23:10:04.902902 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.566572 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:07 crc kubenswrapper[4761]: W1125 23:10:07.574275 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod234edd8e_4c54_43b1_bdf1_b55296bfe91c.slice/crio-872a24ab10085b8d9d34e71e0f65726ce08c1bcabe0ba1234cb6524446329a30 WatchSource:0}: Error finding container 872a24ab10085b8d9d34e71e0f65726ce08c1bcabe0ba1234cb6524446329a30: Status 404 returned error can't find the container with id 872a24ab10085b8d9d34e71e0f65726ce08c1bcabe0ba1234cb6524446329a30 Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.613873 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:07 crc kubenswrapper[4761]: W1125 23:10:07.617225 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ca629d6_a3ee_4125_b410_34c75bb1ba8b.slice/crio-99ab7e4f23130b4ce1e7b505cbf6d34efd30b80709b424a34daffc953ab6422a WatchSource:0}: Error finding container 99ab7e4f23130b4ce1e7b505cbf6d34efd30b80709b424a34daffc953ab6422a: Status 404 returned error can't find the container with id 99ab7e4f23130b4ce1e7b505cbf6d34efd30b80709b424a34daffc953ab6422a Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.895647 4761 generic.go:334] "Generic (PLEG): container finished" podID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerID="ba8dc85743b83428fac971247e2eaad14c53dd9e6f451d9af9a2bf112f7e5c9f" exitCode=0 Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.895750 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerDied","Data":"ba8dc85743b83428fac971247e2eaad14c53dd9e6f451d9af9a2bf112f7e5c9f"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.896130 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerStarted","Data":"872a24ab10085b8d9d34e71e0f65726ce08c1bcabe0ba1234cb6524446329a30"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.898183 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d770ea4c-d11c-4526-af47-7507131606c0","Type":"ContainerStarted","Data":"83a96574ea839cb87fe6cce41919db50c94c3ca651722ca360d99c0dc5af0826"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.900542 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5cf936d6-7508-40de-a1ed-537da8087430","Type":"ContainerStarted","Data":"a635af53fddd95806767fcb2efc69b0b08d925d9cee38f60b0eaa5f6c48b7ead"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.902456 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3c5a3c85-d95b-4252-8feb-155b5b742101","Type":"ContainerStarted","Data":"6d9a8db93e74c877dc1fff20baf6dbd73e41249ddd2b259a3611feea11622828"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.904390 4761 generic.go:334] "Generic (PLEG): container finished" podID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerID="59299f24f6c325d851dbb2020e69259c97da768064216165126665cb058b6ace" exitCode=0 Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.904442 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerDied","Data":"59299f24f6c325d851dbb2020e69259c97da768064216165126665cb058b6ace"} Nov 25 23:10:07 crc kubenswrapper[4761]: I1125 23:10:07.904471 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerStarted","Data":"99ab7e4f23130b4ce1e7b505cbf6d34efd30b80709b424a34daffc953ab6422a"} Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.159947 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.160647 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.162495 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-pxsr2" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.163139 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.169432 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.248275 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-kolla-config\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.248409 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkdcb\" (UniqueName: \"kubernetes.io/projected/a358e604-9896-4108-952a-6ad6560401ab-kube-api-access-zkdcb\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.248448 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-config-data\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.349619 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkdcb\" (UniqueName: \"kubernetes.io/projected/a358e604-9896-4108-952a-6ad6560401ab-kube-api-access-zkdcb\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.349678 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-config-data\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.349728 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-kolla-config\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.350475 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-kolla-config\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.350665 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a358e604-9896-4108-952a-6ad6560401ab-config-data\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.387936 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkdcb\" (UniqueName: \"kubernetes.io/projected/a358e604-9896-4108-952a-6ad6560401ab-kube-api-access-zkdcb\") pod \"memcached-0\" (UID: \"a358e604-9896-4108-952a-6ad6560401ab\") " pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.479903 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:08 crc kubenswrapper[4761]: I1125 23:10:08.921206 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerStarted","Data":"c2d7a273740cbd2484cdff9995f59098a1c285124b3b1509e8bcb558ea9953cf"} Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.043256 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.931359 4761 generic.go:334] "Generic (PLEG): container finished" podID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerID="c2d7a273740cbd2484cdff9995f59098a1c285124b3b1509e8bcb558ea9953cf" exitCode=0 Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.931428 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerDied","Data":"c2d7a273740cbd2484cdff9995f59098a1c285124b3b1509e8bcb558ea9953cf"} Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.932691 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"a358e604-9896-4108-952a-6ad6560401ab","Type":"ContainerStarted","Data":"4116a1a1bfcf6c056db4cb565a1d52958348b9877d74fad8724c40cd52d0c3b5"} Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.935655 4761 generic.go:334] "Generic (PLEG): container finished" podID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerID="513260a7f19bd7b6af4f670e82adf9f8a60c917e7d4089c97c85cc19271a5105" exitCode=0 Nov 25 23:10:09 crc kubenswrapper[4761]: I1125 23:10:09.935716 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerDied","Data":"513260a7f19bd7b6af4f670e82adf9f8a60c917e7d4089c97c85cc19271a5105"} Nov 25 23:10:10 crc kubenswrapper[4761]: I1125 23:10:10.945179 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerStarted","Data":"11451ec4f20a0a81afff3840735c7781bcdbb2693237daf91ad170efbdc87d9b"} Nov 25 23:10:10 crc kubenswrapper[4761]: I1125 23:10:10.947198 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerStarted","Data":"45e984ed018398d176d9df81e46790966fe03a6b9b098c21de469b41694459b1"} Nov 25 23:10:10 crc kubenswrapper[4761]: I1125 23:10:10.974154 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zpjtv" podStartSLOduration=7.513032849 podStartE2EDuration="9.974137925s" podCreationTimestamp="2025-11-25 23:10:01 +0000 UTC" firstStartedPulling="2025-11-25 23:10:07.897131434 +0000 UTC m=+803.630477279" lastFinishedPulling="2025-11-25 23:10:10.35823652 +0000 UTC m=+806.091582355" observedRunningTime="2025-11-25 23:10:10.968366642 +0000 UTC m=+806.701712497" watchObservedRunningTime="2025-11-25 23:10:10.974137925 +0000 UTC m=+806.707483770" Nov 25 23:10:10 crc kubenswrapper[4761]: I1125 23:10:10.989618 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8nmjp" podStartSLOduration=4.598085266 podStartE2EDuration="6.989603717s" podCreationTimestamp="2025-11-25 23:10:04 +0000 UTC" firstStartedPulling="2025-11-25 23:10:07.908642742 +0000 UTC m=+803.641988577" lastFinishedPulling="2025-11-25 23:10:10.300161193 +0000 UTC m=+806.033507028" observedRunningTime="2025-11-25 23:10:10.986931536 +0000 UTC m=+806.720277381" watchObservedRunningTime="2025-11-25 23:10:10.989603717 +0000 UTC m=+806.722949562" Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.916774 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.916822 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.956953 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"a358e604-9896-4108-952a-6ad6560401ab","Type":"ContainerStarted","Data":"8a6c07359d59bd6acb93ff198a85343cd1e15b1e8d9590c85621f34a21b6ce7b"} Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.958167 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.961545 4761 generic.go:334] "Generic (PLEG): container finished" podID="3c5a3c85-d95b-4252-8feb-155b5b742101" containerID="6d9a8db93e74c877dc1fff20baf6dbd73e41249ddd2b259a3611feea11622828" exitCode=0 Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.961585 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3c5a3c85-d95b-4252-8feb-155b5b742101","Type":"ContainerDied","Data":"6d9a8db93e74c877dc1fff20baf6dbd73e41249ddd2b259a3611feea11622828"} Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.964425 4761 generic.go:334] "Generic (PLEG): container finished" podID="d770ea4c-d11c-4526-af47-7507131606c0" containerID="83a96574ea839cb87fe6cce41919db50c94c3ca651722ca360d99c0dc5af0826" exitCode=0 Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.964553 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d770ea4c-d11c-4526-af47-7507131606c0","Type":"ContainerDied","Data":"83a96574ea839cb87fe6cce41919db50c94c3ca651722ca360d99c0dc5af0826"} Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.966964 4761 generic.go:334] "Generic (PLEG): container finished" podID="5cf936d6-7508-40de-a1ed-537da8087430" containerID="a635af53fddd95806767fcb2efc69b0b08d925d9cee38f60b0eaa5f6c48b7ead" exitCode=0 Nov 25 23:10:11 crc kubenswrapper[4761]: I1125 23:10:11.967768 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5cf936d6-7508-40de-a1ed-537da8087430","Type":"ContainerDied","Data":"a635af53fddd95806767fcb2efc69b0b08d925d9cee38f60b0eaa5f6c48b7ead"} Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.001021 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=1.6703248670000002 podStartE2EDuration="4.000992827s" podCreationTimestamp="2025-11-25 23:10:08 +0000 UTC" firstStartedPulling="2025-11-25 23:10:09.04987339 +0000 UTC m=+804.783219225" lastFinishedPulling="2025-11-25 23:10:11.38054135 +0000 UTC m=+807.113887185" observedRunningTime="2025-11-25 23:10:11.987182459 +0000 UTC m=+807.720528324" watchObservedRunningTime="2025-11-25 23:10:12.000992827 +0000 UTC m=+807.734338702" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.578360 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-srnc9"] Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.579829 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.583122 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-p5ttb" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.587162 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-srnc9"] Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.716112 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9sl\" (UniqueName: \"kubernetes.io/projected/43a81c5c-7b19-414c-9e76-9b63aaa0dc14-kube-api-access-zz9sl\") pod \"rabbitmq-cluster-operator-index-srnc9\" (UID: \"43a81c5c-7b19-414c-9e76-9b63aaa0dc14\") " pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.817750 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9sl\" (UniqueName: \"kubernetes.io/projected/43a81c5c-7b19-414c-9e76-9b63aaa0dc14-kube-api-access-zz9sl\") pod \"rabbitmq-cluster-operator-index-srnc9\" (UID: \"43a81c5c-7b19-414c-9e76-9b63aaa0dc14\") " pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.840533 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9sl\" (UniqueName: \"kubernetes.io/projected/43a81c5c-7b19-414c-9e76-9b63aaa0dc14-kube-api-access-zz9sl\") pod \"rabbitmq-cluster-operator-index-srnc9\" (UID: \"43a81c5c-7b19-414c-9e76-9b63aaa0dc14\") " pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.897608 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.981390 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zpjtv" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="registry-server" probeResult="failure" output=< Nov 25 23:10:12 crc kubenswrapper[4761]: timeout: failed to connect service ":50051" within 1s Nov 25 23:10:12 crc kubenswrapper[4761]: > Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.991358 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3c5a3c85-d95b-4252-8feb-155b5b742101","Type":"ContainerStarted","Data":"2275dcb9d676224b88fc886e0b946bab43b1e4a250eaf55f932be33a5becb81e"} Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.996841 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d770ea4c-d11c-4526-af47-7507131606c0","Type":"ContainerStarted","Data":"2b261e789a8a8776609d4337101fff7c77daf04d45a40c2c91c2c567e97a93b9"} Nov 25 23:10:12 crc kubenswrapper[4761]: I1125 23:10:12.999860 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5cf936d6-7508-40de-a1ed-537da8087430","Type":"ContainerStarted","Data":"4fa7123a4d03479ddf988ea87f9461efc46ecab5774fca4a06aaf6efdef58cc2"} Nov 25 23:10:13 crc kubenswrapper[4761]: I1125 23:10:13.014135 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=7.938371223 podStartE2EDuration="16.014113163s" podCreationTimestamp="2025-11-25 23:09:57 +0000 UTC" firstStartedPulling="2025-11-25 23:09:58.98512187 +0000 UTC m=+794.718467705" lastFinishedPulling="2025-11-25 23:10:07.06086381 +0000 UTC m=+802.794209645" observedRunningTime="2025-11-25 23:10:13.012369627 +0000 UTC m=+808.745715542" watchObservedRunningTime="2025-11-25 23:10:13.014113163 +0000 UTC m=+808.747459008" Nov 25 23:10:13 crc kubenswrapper[4761]: I1125 23:10:13.042054 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.115053771 podStartE2EDuration="16.042026117s" podCreationTimestamp="2025-11-25 23:09:57 +0000 UTC" firstStartedPulling="2025-11-25 23:09:59.133948585 +0000 UTC m=+794.867294420" lastFinishedPulling="2025-11-25 23:10:07.060920931 +0000 UTC m=+802.794266766" observedRunningTime="2025-11-25 23:10:13.041273036 +0000 UTC m=+808.774618941" watchObservedRunningTime="2025-11-25 23:10:13.042026117 +0000 UTC m=+808.775371992" Nov 25 23:10:13 crc kubenswrapper[4761]: I1125 23:10:13.077309 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=8.033091698 podStartE2EDuration="16.077295206s" podCreationTimestamp="2025-11-25 23:09:57 +0000 UTC" firstStartedPulling="2025-11-25 23:09:59.091257268 +0000 UTC m=+794.824603103" lastFinishedPulling="2025-11-25 23:10:07.135460776 +0000 UTC m=+802.868806611" observedRunningTime="2025-11-25 23:10:13.074917692 +0000 UTC m=+808.808263547" watchObservedRunningTime="2025-11-25 23:10:13.077295206 +0000 UTC m=+808.810641041" Nov 25 23:10:13 crc kubenswrapper[4761]: I1125 23:10:13.430389 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-srnc9"] Nov 25 23:10:14 crc kubenswrapper[4761]: I1125 23:10:14.008999 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" event={"ID":"43a81c5c-7b19-414c-9e76-9b63aaa0dc14","Type":"ContainerStarted","Data":"1f99013ae14c9b16628f7d53af7611f51b0f44ddd9501608892778e39e3e57b1"} Nov 25 23:10:14 crc kubenswrapper[4761]: I1125 23:10:14.903650 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:14 crc kubenswrapper[4761]: I1125 23:10:14.903974 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:14 crc kubenswrapper[4761]: I1125 23:10:14.970974 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:15 crc kubenswrapper[4761]: I1125 23:10:15.066156 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:15 crc kubenswrapper[4761]: E1125 23:10:15.512955 4761 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.73:52712->38.102.83.73:35801: write tcp 38.102.83.73:52712->38.102.83.73:35801: write: broken pipe Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.035586 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" event={"ID":"43a81c5c-7b19-414c-9e76-9b63aaa0dc14","Type":"ContainerStarted","Data":"64dadcd42c471887956947d4b695abd7c885b5847073dc6ec4044717c6fc0f10"} Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.064028 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" podStartSLOduration=2.84197326 podStartE2EDuration="6.064005324s" podCreationTimestamp="2025-11-25 23:10:12 +0000 UTC" firstStartedPulling="2025-11-25 23:10:13.445301618 +0000 UTC m=+809.178647483" lastFinishedPulling="2025-11-25 23:10:16.667333712 +0000 UTC m=+812.400679547" observedRunningTime="2025-11-25 23:10:18.057269485 +0000 UTC m=+813.790615370" watchObservedRunningTime="2025-11-25 23:10:18.064005324 +0000 UTC m=+813.797351189" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.370521 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.370818 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8nmjp" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="registry-server" containerID="cri-o://45e984ed018398d176d9df81e46790966fe03a6b9b098c21de469b41694459b1" gracePeriod=2 Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.481952 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.580494 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.580539 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.610914 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.611218 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.627536 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:10:18 crc kubenswrapper[4761]: I1125 23:10:18.627583 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:10:19 crc kubenswrapper[4761]: I1125 23:10:19.049256 4761 generic.go:334] "Generic (PLEG): container finished" podID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerID="45e984ed018398d176d9df81e46790966fe03a6b9b098c21de469b41694459b1" exitCode=0 Nov 25 23:10:19 crc kubenswrapper[4761]: I1125 23:10:19.049308 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerDied","Data":"45e984ed018398d176d9df81e46790966fe03a6b9b098c21de469b41694459b1"} Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.400849 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.462627 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content\") pod \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.462828 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmkxl\" (UniqueName: \"kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl\") pod \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.462899 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities\") pod \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\" (UID: \"8ca629d6-a3ee-4125-b410-34c75bb1ba8b\") " Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.464300 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities" (OuterVolumeSpecName: "utilities") pod "8ca629d6-a3ee-4125-b410-34c75bb1ba8b" (UID: "8ca629d6-a3ee-4125-b410-34c75bb1ba8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.475922 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl" (OuterVolumeSpecName: "kube-api-access-qmkxl") pod "8ca629d6-a3ee-4125-b410-34c75bb1ba8b" (UID: "8ca629d6-a3ee-4125-b410-34c75bb1ba8b"). InnerVolumeSpecName "kube-api-access-qmkxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.490829 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ca629d6-a3ee-4125-b410-34c75bb1ba8b" (UID: "8ca629d6-a3ee-4125-b410-34c75bb1ba8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.564764 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.564806 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:21 crc kubenswrapper[4761]: I1125 23:10:21.564821 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmkxl\" (UniqueName: \"kubernetes.io/projected/8ca629d6-a3ee-4125-b410-34c75bb1ba8b-kube-api-access-qmkxl\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:21.999679 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.065338 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.102626 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nmjp" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.102620 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nmjp" event={"ID":"8ca629d6-a3ee-4125-b410-34c75bb1ba8b","Type":"ContainerDied","Data":"99ab7e4f23130b4ce1e7b505cbf6d34efd30b80709b424a34daffc953ab6422a"} Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.102794 4761 scope.go:117] "RemoveContainer" containerID="45e984ed018398d176d9df81e46790966fe03a6b9b098c21de469b41694459b1" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.127561 4761 scope.go:117] "RemoveContainer" containerID="513260a7f19bd7b6af4f670e82adf9f8a60c917e7d4089c97c85cc19271a5105" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.140752 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.148504 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nmjp"] Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.171882 4761 scope.go:117] "RemoveContainer" containerID="59299f24f6c325d851dbb2020e69259c97da768064216165126665cb058b6ace" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.898231 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.898306 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.912356 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:10:22 crc kubenswrapper[4761]: I1125 23:10:22.938233 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:23 crc kubenswrapper[4761]: I1125 23:10:23.018921 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" path="/var/lib/kubelet/pods/8ca629d6-a3ee-4125-b410-34c75bb1ba8b/volumes" Nov 25 23:10:23 crc kubenswrapper[4761]: I1125 23:10:23.030294 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 23:10:23 crc kubenswrapper[4761]: I1125 23:10:23.146236 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-srnc9" Nov 25 23:10:23 crc kubenswrapper[4761]: E1125 23:10:23.380551 4761 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.73:52830->38.102.83.73:35801: write tcp 38.102.83.73:52830->38.102.83.73:35801: write: broken pipe Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.838553 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw"] Nov 25 23:10:25 crc kubenswrapper[4761]: E1125 23:10:25.839342 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="registry-server" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.839365 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="registry-server" Nov 25 23:10:25 crc kubenswrapper[4761]: E1125 23:10:25.839394 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="extract-utilities" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.839407 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="extract-utilities" Nov 25 23:10:25 crc kubenswrapper[4761]: E1125 23:10:25.839457 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="extract-content" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.839473 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="extract-content" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.839680 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ca629d6-a3ee-4125-b410-34c75bb1ba8b" containerName="registry-server" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.841403 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.847069 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.854530 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw"] Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.933970 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.934120 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:25 crc kubenswrapper[4761]: I1125 23:10:25.934180 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljwlr\" (UniqueName: \"kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.035727 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.035861 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljwlr\" (UniqueName: \"kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.035927 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.036407 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.036649 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.059457 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljwlr\" (UniqueName: \"kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.179310 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.484549 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw"] Nov 25 23:10:26 crc kubenswrapper[4761]: W1125 23:10:26.492930 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3986c58_f4d5_430c_a46b_c43b788e0320.slice/crio-74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a WatchSource:0}: Error finding container 74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a: Status 404 returned error can't find the container with id 74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.976623 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:26 crc kubenswrapper[4761]: I1125 23:10:26.980442 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zpjtv" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="registry-server" containerID="cri-o://11451ec4f20a0a81afff3840735c7781bcdbb2693237daf91ad170efbdc87d9b" gracePeriod=2 Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.149118 4761 generic.go:334] "Generic (PLEG): container finished" podID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerID="11451ec4f20a0a81afff3840735c7781bcdbb2693237daf91ad170efbdc87d9b" exitCode=0 Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.149222 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerDied","Data":"11451ec4f20a0a81afff3840735c7781bcdbb2693237daf91ad170efbdc87d9b"} Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.150892 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerID="b15256c3722ad6cafcc0cfc32edf21dd4b0541bd2ef89c1c396b95318deb95e7" exitCode=0 Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.150941 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" event={"ID":"b3986c58-f4d5-430c-a46b-c43b788e0320","Type":"ContainerDied","Data":"b15256c3722ad6cafcc0cfc32edf21dd4b0541bd2ef89c1c396b95318deb95e7"} Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.150974 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" event={"ID":"b3986c58-f4d5-430c-a46b-c43b788e0320","Type":"ContainerStarted","Data":"74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a"} Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.443483 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.451965 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content\") pod \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.452032 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities\") pod \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.452049 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqw9n\" (UniqueName: \"kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n\") pod \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\" (UID: \"234edd8e-4c54-43b1-bdf1-b55296bfe91c\") " Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.453141 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities" (OuterVolumeSpecName: "utilities") pod "234edd8e-4c54-43b1-bdf1-b55296bfe91c" (UID: "234edd8e-4c54-43b1-bdf1-b55296bfe91c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.456662 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n" (OuterVolumeSpecName: "kube-api-access-lqw9n") pod "234edd8e-4c54-43b1-bdf1-b55296bfe91c" (UID: "234edd8e-4c54-43b1-bdf1-b55296bfe91c"). InnerVolumeSpecName "kube-api-access-lqw9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.505368 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "234edd8e-4c54-43b1-bdf1-b55296bfe91c" (UID: "234edd8e-4c54-43b1-bdf1-b55296bfe91c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.553638 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.553667 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/234edd8e-4c54-43b1-bdf1-b55296bfe91c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:27 crc kubenswrapper[4761]: I1125 23:10:27.553677 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqw9n\" (UniqueName: \"kubernetes.io/projected/234edd8e-4c54-43b1-bdf1-b55296bfe91c-kube-api-access-lqw9n\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.158588 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpjtv" event={"ID":"234edd8e-4c54-43b1-bdf1-b55296bfe91c","Type":"ContainerDied","Data":"872a24ab10085b8d9d34e71e0f65726ce08c1bcabe0ba1234cb6524446329a30"} Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.159061 4761 scope.go:117] "RemoveContainer" containerID="11451ec4f20a0a81afff3840735c7781bcdbb2693237daf91ad170efbdc87d9b" Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.158690 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpjtv" Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.177658 4761 scope.go:117] "RemoveContainer" containerID="c2d7a273740cbd2484cdff9995f59098a1c285124b3b1509e8bcb558ea9953cf" Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.196146 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.201887 4761 scope.go:117] "RemoveContainer" containerID="ba8dc85743b83428fac971247e2eaad14c53dd9e6f451d9af9a2bf112f7e5c9f" Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.205535 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zpjtv"] Nov 25 23:10:28 crc kubenswrapper[4761]: I1125 23:10:28.734926 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="d770ea4c-d11c-4526-af47-7507131606c0" containerName="galera" probeResult="failure" output=< Nov 25 23:10:28 crc kubenswrapper[4761]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 25 23:10:28 crc kubenswrapper[4761]: > Nov 25 23:10:29 crc kubenswrapper[4761]: I1125 23:10:29.024066 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" path="/var/lib/kubelet/pods/234edd8e-4c54-43b1-bdf1-b55296bfe91c/volumes" Nov 25 23:10:29 crc kubenswrapper[4761]: I1125 23:10:29.173491 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerID="3d58b882b28e295600d0f6f849c7195d49d17e0d77e3987d0165e61d21f75b7c" exitCode=0 Nov 25 23:10:29 crc kubenswrapper[4761]: I1125 23:10:29.173562 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" event={"ID":"b3986c58-f4d5-430c-a46b-c43b788e0320","Type":"ContainerDied","Data":"3d58b882b28e295600d0f6f849c7195d49d17e0d77e3987d0165e61d21f75b7c"} Nov 25 23:10:30 crc kubenswrapper[4761]: I1125 23:10:30.185754 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerID="b3d13d54d86414be1e7df38527dea53e0669ccfba7c2382da60da389e988b1d0" exitCode=0 Nov 25 23:10:30 crc kubenswrapper[4761]: I1125 23:10:30.185831 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" event={"ID":"b3986c58-f4d5-430c-a46b-c43b788e0320","Type":"ContainerDied","Data":"b3d13d54d86414be1e7df38527dea53e0669ccfba7c2382da60da389e988b1d0"} Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.588637 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.607553 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljwlr\" (UniqueName: \"kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr\") pod \"b3986c58-f4d5-430c-a46b-c43b788e0320\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.607679 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle\") pod \"b3986c58-f4d5-430c-a46b-c43b788e0320\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.607875 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util\") pod \"b3986c58-f4d5-430c-a46b-c43b788e0320\" (UID: \"b3986c58-f4d5-430c-a46b-c43b788e0320\") " Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.608683 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle" (OuterVolumeSpecName: "bundle") pod "b3986c58-f4d5-430c-a46b-c43b788e0320" (UID: "b3986c58-f4d5-430c-a46b-c43b788e0320"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.612531 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr" (OuterVolumeSpecName: "kube-api-access-ljwlr") pod "b3986c58-f4d5-430c-a46b-c43b788e0320" (UID: "b3986c58-f4d5-430c-a46b-c43b788e0320"). InnerVolumeSpecName "kube-api-access-ljwlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.626732 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util" (OuterVolumeSpecName: "util") pod "b3986c58-f4d5-430c-a46b-c43b788e0320" (UID: "b3986c58-f4d5-430c-a46b-c43b788e0320"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.709142 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.709182 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljwlr\" (UniqueName: \"kubernetes.io/projected/b3986c58-f4d5-430c-a46b-c43b788e0320-kube-api-access-ljwlr\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.709193 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b3986c58-f4d5-430c-a46b-c43b788e0320-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.978476 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979133 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="util" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979149 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="util" Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979166 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="extract-utilities" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979175 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="extract-utilities" Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979195 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="pull" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979203 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="pull" Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979212 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="extract" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979219 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="extract" Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979234 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="registry-server" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979243 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="registry-server" Nov 25 23:10:31 crc kubenswrapper[4761]: E1125 23:10:31.979254 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="extract-content" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979262 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="extract-content" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979394 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3986c58-f4d5-430c-a46b-c43b788e0320" containerName="extract" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.979405 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="234edd8e-4c54-43b1-bdf1-b55296bfe91c" containerName="registry-server" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.980368 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:31 crc kubenswrapper[4761]: I1125 23:10:31.999478 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.024788 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.024907 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98wpk\" (UniqueName: \"kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.024948 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.126289 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.126396 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98wpk\" (UniqueName: \"kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.126421 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.126875 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.126980 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.163527 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98wpk\" (UniqueName: \"kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk\") pod \"certified-operators-kfgvw\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.202617 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" event={"ID":"b3986c58-f4d5-430c-a46b-c43b788e0320","Type":"ContainerDied","Data":"74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a"} Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.202661 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74653294a28bccbbbde41d72bdc5525be7575853c52b1484fbfa7a8449e4c54a" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.202671 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.308793 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:32 crc kubenswrapper[4761]: I1125 23:10:32.762741 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:33 crc kubenswrapper[4761]: I1125 23:10:33.214177 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerID="c39fdd93ed4dec447d47790f635529e834536236ac0c0cb1eed48d2550b561ea" exitCode=0 Nov 25 23:10:33 crc kubenswrapper[4761]: I1125 23:10:33.214239 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerDied","Data":"c39fdd93ed4dec447d47790f635529e834536236ac0c0cb1eed48d2550b561ea"} Nov 25 23:10:33 crc kubenswrapper[4761]: I1125 23:10:33.214281 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerStarted","Data":"7fd87051b964ef5823a7bee5574de0592126a328d1fb053a945e49958976584e"} Nov 25 23:10:34 crc kubenswrapper[4761]: I1125 23:10:34.224154 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerStarted","Data":"766ba2aa969e4a0497e16c2b440787fbc7a569d2912a7796704cdbe4099f847c"} Nov 25 23:10:34 crc kubenswrapper[4761]: I1125 23:10:34.841545 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:10:34 crc kubenswrapper[4761]: I1125 23:10:34.963969 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 23:10:35 crc kubenswrapper[4761]: I1125 23:10:35.242328 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerID="766ba2aa969e4a0497e16c2b440787fbc7a569d2912a7796704cdbe4099f847c" exitCode=0 Nov 25 23:10:35 crc kubenswrapper[4761]: I1125 23:10:35.243755 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerDied","Data":"766ba2aa969e4a0497e16c2b440787fbc7a569d2912a7796704cdbe4099f847c"} Nov 25 23:10:35 crc kubenswrapper[4761]: I1125 23:10:35.244788 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerStarted","Data":"3a1c27674afd36282dcc26ea1aabd8cc281f581c824b16ea435e22f1c24509fc"} Nov 25 23:10:35 crc kubenswrapper[4761]: I1125 23:10:35.271327 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfgvw" podStartSLOduration=2.62750428 podStartE2EDuration="4.271309216s" podCreationTimestamp="2025-11-25 23:10:31 +0000 UTC" firstStartedPulling="2025-11-25 23:10:33.217053577 +0000 UTC m=+828.950399442" lastFinishedPulling="2025-11-25 23:10:34.860858503 +0000 UTC m=+830.594204378" observedRunningTime="2025-11-25 23:10:35.267185076 +0000 UTC m=+831.000530921" watchObservedRunningTime="2025-11-25 23:10:35.271309216 +0000 UTC m=+831.004655061" Nov 25 23:10:37 crc kubenswrapper[4761]: I1125 23:10:37.274308 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:10:37 crc kubenswrapper[4761]: I1125 23:10:37.338230 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.774263 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.775450 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.784916 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.918539 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfq7l\" (UniqueName: \"kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.918784 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:38 crc kubenswrapper[4761]: I1125 23:10:38.918869 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.020377 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.020457 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.020572 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfq7l\" (UniqueName: \"kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.021201 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.021223 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.052793 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfq7l\" (UniqueName: \"kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l\") pod \"redhat-operators-c6dsw\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.090499 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:39 crc kubenswrapper[4761]: I1125 23:10:39.402516 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:40 crc kubenswrapper[4761]: I1125 23:10:40.296256 4761 generic.go:334] "Generic (PLEG): container finished" podID="5f584074-1ece-492a-8a10-530dca6b313e" containerID="aa1e591ea871acfeb9dbcb3b38cdb21592dbf5bef66fa70fad6ec741e69c73ba" exitCode=0 Nov 25 23:10:40 crc kubenswrapper[4761]: I1125 23:10:40.296311 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerDied","Data":"aa1e591ea871acfeb9dbcb3b38cdb21592dbf5bef66fa70fad6ec741e69c73ba"} Nov 25 23:10:40 crc kubenswrapper[4761]: I1125 23:10:40.296345 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerStarted","Data":"3b6f5076d57bf5527502a10c93ff5491ec6d53e59f7385e27d8be37b1285ea9d"} Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.305808 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerStarted","Data":"aebddb0772e80de849a6b8e7591fbae656eb6acd01cad216b76869ccde127c06"} Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.587466 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg"] Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.588830 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.600277 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-752bb" Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.606283 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg"] Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.756338 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r5cx\" (UniqueName: \"kubernetes.io/projected/91b5b48e-a642-45e2-a84a-d449123e57bb-kube-api-access-8r5cx\") pod \"rabbitmq-cluster-operator-779fc9694b-cskvg\" (UID: \"91b5b48e-a642-45e2-a84a-d449123e57bb\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.858582 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r5cx\" (UniqueName: \"kubernetes.io/projected/91b5b48e-a642-45e2-a84a-d449123e57bb-kube-api-access-8r5cx\") pod \"rabbitmq-cluster-operator-779fc9694b-cskvg\" (UID: \"91b5b48e-a642-45e2-a84a-d449123e57bb\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.880663 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r5cx\" (UniqueName: \"kubernetes.io/projected/91b5b48e-a642-45e2-a84a-d449123e57bb-kube-api-access-8r5cx\") pod \"rabbitmq-cluster-operator-779fc9694b-cskvg\" (UID: \"91b5b48e-a642-45e2-a84a-d449123e57bb\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" Nov 25 23:10:41 crc kubenswrapper[4761]: I1125 23:10:41.909585 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.309729 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.310025 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.318732 4761 generic.go:334] "Generic (PLEG): container finished" podID="5f584074-1ece-492a-8a10-530dca6b313e" containerID="aebddb0772e80de849a6b8e7591fbae656eb6acd01cad216b76869ccde127c06" exitCode=0 Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.318797 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerDied","Data":"aebddb0772e80de849a6b8e7591fbae656eb6acd01cad216b76869ccde127c06"} Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.375455 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:42 crc kubenswrapper[4761]: I1125 23:10:42.413099 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg"] Nov 25 23:10:43 crc kubenswrapper[4761]: I1125 23:10:43.327999 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerStarted","Data":"f8a2946a717553fd8a698f331357146d8b03085b2fb0ac33101b6ff92fecbcad"} Nov 25 23:10:43 crc kubenswrapper[4761]: I1125 23:10:43.399777 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:44 crc kubenswrapper[4761]: I1125 23:10:44.343935 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerStarted","Data":"dee01c4d40a2c477df8b328d39c4e5369257d1369f41d5322168145e10cd54c6"} Nov 25 23:10:44 crc kubenswrapper[4761]: I1125 23:10:44.365357 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6dsw" podStartSLOduration=3.375623403 podStartE2EDuration="6.365336718s" podCreationTimestamp="2025-11-25 23:10:38 +0000 UTC" firstStartedPulling="2025-11-25 23:10:40.298190175 +0000 UTC m=+836.031536010" lastFinishedPulling="2025-11-25 23:10:43.28790346 +0000 UTC m=+839.021249325" observedRunningTime="2025-11-25 23:10:44.364097395 +0000 UTC m=+840.097443280" watchObservedRunningTime="2025-11-25 23:10:44.365336718 +0000 UTC m=+840.098682563" Nov 25 23:10:47 crc kubenswrapper[4761]: I1125 23:10:47.363763 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerStarted","Data":"bf334415f6cd4cfd9d90cdd581fa39798b88c952ac303a041dae5a20705e849e"} Nov 25 23:10:47 crc kubenswrapper[4761]: I1125 23:10:47.380212 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" podStartSLOduration=2.121068085 podStartE2EDuration="6.380197644s" podCreationTimestamp="2025-11-25 23:10:41 +0000 UTC" firstStartedPulling="2025-11-25 23:10:42.424267975 +0000 UTC m=+838.157613810" lastFinishedPulling="2025-11-25 23:10:46.683397534 +0000 UTC m=+842.416743369" observedRunningTime="2025-11-25 23:10:47.379198477 +0000 UTC m=+843.112544392" watchObservedRunningTime="2025-11-25 23:10:47.380197644 +0000 UTC m=+843.113543479" Nov 25 23:10:47 crc kubenswrapper[4761]: I1125 23:10:47.778075 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:47 crc kubenswrapper[4761]: I1125 23:10:47.778986 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfgvw" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="registry-server" containerID="cri-o://3a1c27674afd36282dcc26ea1aabd8cc281f581c824b16ea435e22f1c24509fc" gracePeriod=2 Nov 25 23:10:48 crc kubenswrapper[4761]: I1125 23:10:48.372862 4761 generic.go:334] "Generic (PLEG): container finished" podID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerID="3a1c27674afd36282dcc26ea1aabd8cc281f581c824b16ea435e22f1c24509fc" exitCode=0 Nov 25 23:10:48 crc kubenswrapper[4761]: I1125 23:10:48.372961 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerDied","Data":"3a1c27674afd36282dcc26ea1aabd8cc281f581c824b16ea435e22f1c24509fc"} Nov 25 23:10:48 crc kubenswrapper[4761]: I1125 23:10:48.997820 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.091853 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.091895 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.142161 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.184637 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content\") pod \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.184792 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98wpk\" (UniqueName: \"kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk\") pod \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.184858 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities\") pod \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\" (UID: \"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6\") " Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.185679 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities" (OuterVolumeSpecName: "utilities") pod "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" (UID: "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.200020 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk" (OuterVolumeSpecName: "kube-api-access-98wpk") pod "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" (UID: "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6"). InnerVolumeSpecName "kube-api-access-98wpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.235813 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" (UID: "b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.287604 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98wpk\" (UniqueName: \"kubernetes.io/projected/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-kube-api-access-98wpk\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.288127 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.288293 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.382935 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfgvw" event={"ID":"b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6","Type":"ContainerDied","Data":"7fd87051b964ef5823a7bee5574de0592126a328d1fb053a945e49958976584e"} Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.383013 4761 scope.go:117] "RemoveContainer" containerID="3a1c27674afd36282dcc26ea1aabd8cc281f581c824b16ea435e22f1c24509fc" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.387965 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfgvw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.403369 4761 scope.go:117] "RemoveContainer" containerID="766ba2aa969e4a0497e16c2b440787fbc7a569d2912a7796704cdbe4099f847c" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.441247 4761 scope.go:117] "RemoveContainer" containerID="c39fdd93ed4dec447d47790f635529e834536236ac0c0cb1eed48d2550b561ea" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.510705 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.511043 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:49 crc kubenswrapper[4761]: I1125 23:10:49.513606 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfgvw"] Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.582279 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 23:10:50 crc kubenswrapper[4761]: E1125 23:10:50.582750 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="extract-utilities" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.582776 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="extract-utilities" Nov 25 23:10:50 crc kubenswrapper[4761]: E1125 23:10:50.582793 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="registry-server" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.582806 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="registry-server" Nov 25 23:10:50 crc kubenswrapper[4761]: E1125 23:10:50.582842 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="extract-content" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.582855 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="extract-content" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.583095 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" containerName="registry-server" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.584255 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.587255 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.588619 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.588804 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-cfkgt" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.589089 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.589473 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.606287 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707638 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4370dc6c-2ed4-425a-bc57-98c19c37fde9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707763 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4gsb\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-kube-api-access-x4gsb\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707802 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707825 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4370dc6c-2ed4-425a-bc57-98c19c37fde9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707859 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707882 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.707973 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.708052 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4370dc6c-2ed4-425a-bc57-98c19c37fde9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809606 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4gsb\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-kube-api-access-x4gsb\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809735 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809785 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4370dc6c-2ed4-425a-bc57-98c19c37fde9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809845 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809881 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809915 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.809960 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4370dc6c-2ed4-425a-bc57-98c19c37fde9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.810013 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4370dc6c-2ed4-425a-bc57-98c19c37fde9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.811513 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4370dc6c-2ed4-425a-bc57-98c19c37fde9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.814023 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.814540 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.821795 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4370dc6c-2ed4-425a-bc57-98c19c37fde9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.821871 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.822412 4761 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.822458 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/14a863c5273b1967b4c05f43e97bb17c6e6314f1efed8017cd0e8459fcedf800/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.822626 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4370dc6c-2ed4-425a-bc57-98c19c37fde9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.847904 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4gsb\" (UniqueName: \"kubernetes.io/projected/4370dc6c-2ed4-425a-bc57-98c19c37fde9-kube-api-access-x4gsb\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.872968 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61b94d3d-6b1f-4549-93d5-0cca2898029d\") pod \"rabbitmq-server-0\" (UID: \"4370dc6c-2ed4-425a-bc57-98c19c37fde9\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:50 crc kubenswrapper[4761]: I1125 23:10:50.923636 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:10:51 crc kubenswrapper[4761]: I1125 23:10:51.023891 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6" path="/var/lib/kubelet/pods/b3c8cd9c-e376-40c3-abbc-bf3eb9e294b6/volumes" Nov 25 23:10:51 crc kubenswrapper[4761]: I1125 23:10:51.287580 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 23:10:51 crc kubenswrapper[4761]: I1125 23:10:51.400800 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4370dc6c-2ed4-425a-bc57-98c19c37fde9","Type":"ContainerStarted","Data":"05600ac8199666deaa68be6df6fde70a37970766a0826d86eb8aa139519f7359"} Nov 25 23:10:52 crc kubenswrapper[4761]: I1125 23:10:52.770904 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:52 crc kubenswrapper[4761]: I1125 23:10:52.771304 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6dsw" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="registry-server" containerID="cri-o://dee01c4d40a2c477df8b328d39c4e5369257d1369f41d5322168145e10cd54c6" gracePeriod=2 Nov 25 23:10:53 crc kubenswrapper[4761]: I1125 23:10:53.427757 4761 generic.go:334] "Generic (PLEG): container finished" podID="5f584074-1ece-492a-8a10-530dca6b313e" containerID="dee01c4d40a2c477df8b328d39c4e5369257d1369f41d5322168145e10cd54c6" exitCode=0 Nov 25 23:10:53 crc kubenswrapper[4761]: I1125 23:10:53.427832 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerDied","Data":"dee01c4d40a2c477df8b328d39c4e5369257d1369f41d5322168145e10cd54c6"} Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.045571 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.176246 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities\") pod \"5f584074-1ece-492a-8a10-530dca6b313e\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.176294 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfq7l\" (UniqueName: \"kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l\") pod \"5f584074-1ece-492a-8a10-530dca6b313e\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.176318 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content\") pod \"5f584074-1ece-492a-8a10-530dca6b313e\" (UID: \"5f584074-1ece-492a-8a10-530dca6b313e\") " Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.177285 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities" (OuterVolumeSpecName: "utilities") pod "5f584074-1ece-492a-8a10-530dca6b313e" (UID: "5f584074-1ece-492a-8a10-530dca6b313e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.182649 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l" (OuterVolumeSpecName: "kube-api-access-kfq7l") pod "5f584074-1ece-492a-8a10-530dca6b313e" (UID: "5f584074-1ece-492a-8a10-530dca6b313e"). InnerVolumeSpecName "kube-api-access-kfq7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.270178 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f584074-1ece-492a-8a10-530dca6b313e" (UID: "5f584074-1ece-492a-8a10-530dca6b313e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.278340 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.278381 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfq7l\" (UniqueName: \"kubernetes.io/projected/5f584074-1ece-492a-8a10-530dca6b313e-kube-api-access-kfq7l\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.278393 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f584074-1ece-492a-8a10-530dca6b313e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.448924 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dsw" event={"ID":"5f584074-1ece-492a-8a10-530dca6b313e","Type":"ContainerDied","Data":"3b6f5076d57bf5527502a10c93ff5491ec6d53e59f7385e27d8be37b1285ea9d"} Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.448997 4761 scope.go:117] "RemoveContainer" containerID="dee01c4d40a2c477df8b328d39c4e5369257d1369f41d5322168145e10cd54c6" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.449036 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dsw" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.490583 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.502664 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6dsw"] Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.778719 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-sw22g"] Nov 25 23:10:55 crc kubenswrapper[4761]: E1125 23:10:55.778996 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="registry-server" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.779020 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="registry-server" Nov 25 23:10:55 crc kubenswrapper[4761]: E1125 23:10:55.779038 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="extract-utilities" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.779046 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="extract-utilities" Nov 25 23:10:55 crc kubenswrapper[4761]: E1125 23:10:55.779060 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="extract-content" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.779069 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="extract-content" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.779187 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f584074-1ece-492a-8a10-530dca6b313e" containerName="registry-server" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.779575 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.783970 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-n6xdd" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.787986 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-sw22g"] Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.889515 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2v2h\" (UniqueName: \"kubernetes.io/projected/464d8029-30b2-4ac1-8820-93dd0f144047-kube-api-access-p2v2h\") pod \"keystone-operator-index-sw22g\" (UID: \"464d8029-30b2-4ac1-8820-93dd0f144047\") " pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:10:55 crc kubenswrapper[4761]: I1125 23:10:55.991924 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2v2h\" (UniqueName: \"kubernetes.io/projected/464d8029-30b2-4ac1-8820-93dd0f144047-kube-api-access-p2v2h\") pod \"keystone-operator-index-sw22g\" (UID: \"464d8029-30b2-4ac1-8820-93dd0f144047\") " pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:10:56 crc kubenswrapper[4761]: I1125 23:10:56.008028 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2v2h\" (UniqueName: \"kubernetes.io/projected/464d8029-30b2-4ac1-8820-93dd0f144047-kube-api-access-p2v2h\") pod \"keystone-operator-index-sw22g\" (UID: \"464d8029-30b2-4ac1-8820-93dd0f144047\") " pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:10:56 crc kubenswrapper[4761]: I1125 23:10:56.107020 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:10:56 crc kubenswrapper[4761]: I1125 23:10:56.229051 4761 scope.go:117] "RemoveContainer" containerID="aebddb0772e80de849a6b8e7591fbae656eb6acd01cad216b76869ccde127c06" Nov 25 23:10:56 crc kubenswrapper[4761]: I1125 23:10:56.298575 4761 scope.go:117] "RemoveContainer" containerID="aa1e591ea871acfeb9dbcb3b38cdb21592dbf5bef66fa70fad6ec741e69c73ba" Nov 25 23:10:57 crc kubenswrapper[4761]: I1125 23:10:57.023990 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f584074-1ece-492a-8a10-530dca6b313e" path="/var/lib/kubelet/pods/5f584074-1ece-492a-8a10-530dca6b313e/volumes" Nov 25 23:10:57 crc kubenswrapper[4761]: I1125 23:10:57.438545 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-sw22g"] Nov 25 23:10:57 crc kubenswrapper[4761]: W1125 23:10:57.441569 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod464d8029_30b2_4ac1_8820_93dd0f144047.slice/crio-a8a61c11227c19ef32656daaa40a2602cdbfd01428b19d9cd8ac42036bdb1a01 WatchSource:0}: Error finding container a8a61c11227c19ef32656daaa40a2602cdbfd01428b19d9cd8ac42036bdb1a01: Status 404 returned error can't find the container with id a8a61c11227c19ef32656daaa40a2602cdbfd01428b19d9cd8ac42036bdb1a01 Nov 25 23:10:57 crc kubenswrapper[4761]: I1125 23:10:57.462048 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-sw22g" event={"ID":"464d8029-30b2-4ac1-8820-93dd0f144047","Type":"ContainerStarted","Data":"a8a61c11227c19ef32656daaa40a2602cdbfd01428b19d9cd8ac42036bdb1a01"} Nov 25 23:10:58 crc kubenswrapper[4761]: I1125 23:10:58.472798 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-sw22g" event={"ID":"464d8029-30b2-4ac1-8820-93dd0f144047","Type":"ContainerStarted","Data":"0b7153fdd578b82b968c5ae086c9f1fbd227be7c984963fa0d9532f1059ce652"} Nov 25 23:10:58 crc kubenswrapper[4761]: I1125 23:10:58.475297 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4370dc6c-2ed4-425a-bc57-98c19c37fde9","Type":"ContainerStarted","Data":"bb57c39b2b1e161f1b758eddf3e0821503dc6d601972ea33789dfff83380959c"} Nov 25 23:10:58 crc kubenswrapper[4761]: I1125 23:10:58.500874 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-sw22g" podStartSLOduration=2.685076469 podStartE2EDuration="3.500847728s" podCreationTimestamp="2025-11-25 23:10:55 +0000 UTC" firstStartedPulling="2025-11-25 23:10:57.443412052 +0000 UTC m=+853.176757877" lastFinishedPulling="2025-11-25 23:10:58.259183261 +0000 UTC m=+853.992529136" observedRunningTime="2025-11-25 23:10:58.497291324 +0000 UTC m=+854.230637219" watchObservedRunningTime="2025-11-25 23:10:58.500847728 +0000 UTC m=+854.234193593" Nov 25 23:11:06 crc kubenswrapper[4761]: I1125 23:11:06.108065 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:11:06 crc kubenswrapper[4761]: I1125 23:11:06.108841 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:11:06 crc kubenswrapper[4761]: I1125 23:11:06.144766 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:11:06 crc kubenswrapper[4761]: I1125 23:11:06.570689 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-sw22g" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.659024 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q"] Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.660571 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.665434 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.683360 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q"] Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.788925 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.788981 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.789165 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75wf2\" (UniqueName: \"kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.890379 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.890442 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.890488 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75wf2\" (UniqueName: \"kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.891047 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.891250 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.922641 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75wf2\" (UniqueName: \"kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:08 crc kubenswrapper[4761]: I1125 23:11:08.990050 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:09 crc kubenswrapper[4761]: I1125 23:11:09.292024 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q"] Nov 25 23:11:09 crc kubenswrapper[4761]: W1125 23:11:09.298674 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1206534d_9218_4636_b769_46aab59ed934.slice/crio-9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620 WatchSource:0}: Error finding container 9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620: Status 404 returned error can't find the container with id 9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620 Nov 25 23:11:09 crc kubenswrapper[4761]: I1125 23:11:09.564779 4761 generic.go:334] "Generic (PLEG): container finished" podID="1206534d-9218-4636-b769-46aab59ed934" containerID="21d4970dc9e79137da412142fa8b85035ed87ff6c37179bf80d75b48937f2d8c" exitCode=0 Nov 25 23:11:09 crc kubenswrapper[4761]: I1125 23:11:09.564842 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" event={"ID":"1206534d-9218-4636-b769-46aab59ed934","Type":"ContainerDied","Data":"21d4970dc9e79137da412142fa8b85035ed87ff6c37179bf80d75b48937f2d8c"} Nov 25 23:11:09 crc kubenswrapper[4761]: I1125 23:11:09.564881 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" event={"ID":"1206534d-9218-4636-b769-46aab59ed934","Type":"ContainerStarted","Data":"9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620"} Nov 25 23:11:10 crc kubenswrapper[4761]: I1125 23:11:10.573973 4761 generic.go:334] "Generic (PLEG): container finished" podID="1206534d-9218-4636-b769-46aab59ed934" containerID="22786e932583a444c20f1c737d27f325e22b4be6fcf20e05b4ccb4eefb5c3e08" exitCode=0 Nov 25 23:11:10 crc kubenswrapper[4761]: I1125 23:11:10.574160 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" event={"ID":"1206534d-9218-4636-b769-46aab59ed934","Type":"ContainerDied","Data":"22786e932583a444c20f1c737d27f325e22b4be6fcf20e05b4ccb4eefb5c3e08"} Nov 25 23:11:11 crc kubenswrapper[4761]: I1125 23:11:11.585541 4761 generic.go:334] "Generic (PLEG): container finished" podID="1206534d-9218-4636-b769-46aab59ed934" containerID="567dbf2a568ccf93baaf59396f74624983dde94cd5f01a3d0d9214523ecbff0f" exitCode=0 Nov 25 23:11:11 crc kubenswrapper[4761]: I1125 23:11:11.585694 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" event={"ID":"1206534d-9218-4636-b769-46aab59ed934","Type":"ContainerDied","Data":"567dbf2a568ccf93baaf59396f74624983dde94cd5f01a3d0d9214523ecbff0f"} Nov 25 23:11:12 crc kubenswrapper[4761]: I1125 23:11:12.996912 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.156983 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util\") pod \"1206534d-9218-4636-b769-46aab59ed934\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.157094 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75wf2\" (UniqueName: \"kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2\") pod \"1206534d-9218-4636-b769-46aab59ed934\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.157124 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle\") pod \"1206534d-9218-4636-b769-46aab59ed934\" (UID: \"1206534d-9218-4636-b769-46aab59ed934\") " Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.158633 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle" (OuterVolumeSpecName: "bundle") pod "1206534d-9218-4636-b769-46aab59ed934" (UID: "1206534d-9218-4636-b769-46aab59ed934"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.166503 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2" (OuterVolumeSpecName: "kube-api-access-75wf2") pod "1206534d-9218-4636-b769-46aab59ed934" (UID: "1206534d-9218-4636-b769-46aab59ed934"). InnerVolumeSpecName "kube-api-access-75wf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.172565 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util" (OuterVolumeSpecName: "util") pod "1206534d-9218-4636-b769-46aab59ed934" (UID: "1206534d-9218-4636-b769-46aab59ed934"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.258551 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.258587 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1206534d-9218-4636-b769-46aab59ed934-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.258604 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75wf2\" (UniqueName: \"kubernetes.io/projected/1206534d-9218-4636-b769-46aab59ed934-kube-api-access-75wf2\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.603262 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" event={"ID":"1206534d-9218-4636-b769-46aab59ed934","Type":"ContainerDied","Data":"9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620"} Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.603595 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ab3e40e15fde6c2f7b8f24f6240d5a8a5c383ae8f29007e44f4837f64e1e620" Nov 25 23:11:13 crc kubenswrapper[4761]: I1125 23:11:13.603321 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.517110 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv"] Nov 25 23:11:24 crc kubenswrapper[4761]: E1125 23:11:24.518009 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="util" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.518029 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="util" Nov 25 23:11:24 crc kubenswrapper[4761]: E1125 23:11:24.518061 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="pull" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.518072 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="pull" Nov 25 23:11:24 crc kubenswrapper[4761]: E1125 23:11:24.518088 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="extract" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.518100 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="extract" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.518284 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="1206534d-9218-4636-b769-46aab59ed934" containerName="extract" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.518902 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.525150 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.525151 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-msgsj" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.552399 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv"] Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.624679 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-apiservice-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.624758 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-webhook-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.624886 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvf4\" (UniqueName: \"kubernetes.io/projected/47a4708b-be53-4783-b9c1-9536d90b9e4c-kube-api-access-bhvf4\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.726361 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvf4\" (UniqueName: \"kubernetes.io/projected/47a4708b-be53-4783-b9c1-9536d90b9e4c-kube-api-access-bhvf4\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.726441 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-apiservice-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.726473 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-webhook-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.731887 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-apiservice-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.732457 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47a4708b-be53-4783-b9c1-9536d90b9e4c-webhook-cert\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.742476 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvf4\" (UniqueName: \"kubernetes.io/projected/47a4708b-be53-4783-b9c1-9536d90b9e4c-kube-api-access-bhvf4\") pod \"keystone-operator-controller-manager-8559b887bb-frnqv\" (UID: \"47a4708b-be53-4783-b9c1-9536d90b9e4c\") " pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:24 crc kubenswrapper[4761]: I1125 23:11:24.840640 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:25 crc kubenswrapper[4761]: I1125 23:11:25.595027 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv"] Nov 25 23:11:25 crc kubenswrapper[4761]: W1125 23:11:25.603845 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47a4708b_be53_4783_b9c1_9536d90b9e4c.slice/crio-d2aaa05918b5917374f3d8f566b91f60ab42a6bf4f6904930f58ed735cc25ee5 WatchSource:0}: Error finding container d2aaa05918b5917374f3d8f566b91f60ab42a6bf4f6904930f58ed735cc25ee5: Status 404 returned error can't find the container with id d2aaa05918b5917374f3d8f566b91f60ab42a6bf4f6904930f58ed735cc25ee5 Nov 25 23:11:25 crc kubenswrapper[4761]: I1125 23:11:25.710516 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerStarted","Data":"d2aaa05918b5917374f3d8f566b91f60ab42a6bf4f6904930f58ed735cc25ee5"} Nov 25 23:11:29 crc kubenswrapper[4761]: I1125 23:11:29.734972 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerStarted","Data":"cde14783c38bcdf6cfb028f2cebbd11e85d7de90cd9950c41d54501602a21e86"} Nov 25 23:11:29 crc kubenswrapper[4761]: I1125 23:11:29.735611 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:30 crc kubenswrapper[4761]: I1125 23:11:30.743661 4761 generic.go:334] "Generic (PLEG): container finished" podID="4370dc6c-2ed4-425a-bc57-98c19c37fde9" containerID="bb57c39b2b1e161f1b758eddf3e0821503dc6d601972ea33789dfff83380959c" exitCode=0 Nov 25 23:11:30 crc kubenswrapper[4761]: I1125 23:11:30.743771 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4370dc6c-2ed4-425a-bc57-98c19c37fde9","Type":"ContainerDied","Data":"bb57c39b2b1e161f1b758eddf3e0821503dc6d601972ea33789dfff83380959c"} Nov 25 23:11:30 crc kubenswrapper[4761]: I1125 23:11:30.774959 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podStartSLOduration=3.300011356 podStartE2EDuration="6.774934345s" podCreationTimestamp="2025-11-25 23:11:24 +0000 UTC" firstStartedPulling="2025-11-25 23:11:25.606397094 +0000 UTC m=+881.339742979" lastFinishedPulling="2025-11-25 23:11:29.081320123 +0000 UTC m=+884.814665968" observedRunningTime="2025-11-25 23:11:29.771255761 +0000 UTC m=+885.504601626" watchObservedRunningTime="2025-11-25 23:11:30.774934345 +0000 UTC m=+886.508280190" Nov 25 23:11:31 crc kubenswrapper[4761]: I1125 23:11:31.755541 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4370dc6c-2ed4-425a-bc57-98c19c37fde9","Type":"ContainerStarted","Data":"e0ae37bd9902dae29e0d74b89634a95edc87b6f82f4b59e97a56b85181f0bce6"} Nov 25 23:11:31 crc kubenswrapper[4761]: I1125 23:11:31.756071 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:11:31 crc kubenswrapper[4761]: I1125 23:11:31.788649 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.993799184 podStartE2EDuration="42.788633347s" podCreationTimestamp="2025-11-25 23:10:49 +0000 UTC" firstStartedPulling="2025-11-25 23:10:51.294689772 +0000 UTC m=+847.028035607" lastFinishedPulling="2025-11-25 23:10:57.089523935 +0000 UTC m=+852.822869770" observedRunningTime="2025-11-25 23:11:31.787096626 +0000 UTC m=+887.520442481" watchObservedRunningTime="2025-11-25 23:11:31.788633347 +0000 UTC m=+887.521979182" Nov 25 23:11:34 crc kubenswrapper[4761]: I1125 23:11:34.847409 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.075083 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-hdrz6"] Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.077092 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.080535 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8"] Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.082293 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.084310 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.086041 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hdrz6"] Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.099805 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8"] Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.260756 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.260854 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.260902 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqd86\" (UniqueName: \"kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.260965 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrpx\" (UniqueName: \"kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.362863 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.362943 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.362977 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqd86\" (UniqueName: \"kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.363026 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrpx\" (UniqueName: \"kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.364107 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.364795 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.403363 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqd86\" (UniqueName: \"kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86\") pod \"keystone-db-create-hdrz6\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.407862 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrpx\" (UniqueName: \"kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx\") pod \"keystone-edb2-account-create-update-cn8n8\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.417609 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.698600 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.867642 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8"] Nov 25 23:11:40 crc kubenswrapper[4761]: I1125 23:11:40.947953 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 23:11:40 crc kubenswrapper[4761]: W1125 23:11:40.998071 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32d39ec9_e2eb_41a4_83d8_f0f84fb3681f.slice/crio-0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9 WatchSource:0}: Error finding container 0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9: Status 404 returned error can't find the container with id 0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9 Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.004639 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hdrz6"] Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.846844 4761 generic.go:334] "Generic (PLEG): container finished" podID="cda8b5ee-b6f4-4eb2-a095-f9647da330f6" containerID="9ceee44d6828c43af4df267956dcd49e8cddaca7d565adff59bc4070c472a393" exitCode=0 Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.846902 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" event={"ID":"cda8b5ee-b6f4-4eb2-a095-f9647da330f6","Type":"ContainerDied","Data":"9ceee44d6828c43af4df267956dcd49e8cddaca7d565adff59bc4070c472a393"} Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.847409 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" event={"ID":"cda8b5ee-b6f4-4eb2-a095-f9647da330f6","Type":"ContainerStarted","Data":"e515ccb3c4f8e11fb787cea8cb9c7ec7af93cb3e2b112bd613be491885ff4697"} Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.850133 4761 generic.go:334] "Generic (PLEG): container finished" podID="32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" containerID="e5eeb6a02324099e65ff127469671ffbc53b2b026e5594dd23f3b394d041beeb" exitCode=0 Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.850224 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hdrz6" event={"ID":"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f","Type":"ContainerDied","Data":"e5eeb6a02324099e65ff127469671ffbc53b2b026e5594dd23f3b394d041beeb"} Nov 25 23:11:41 crc kubenswrapper[4761]: I1125 23:11:41.850259 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hdrz6" event={"ID":"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f","Type":"ContainerStarted","Data":"0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9"} Nov 25 23:11:42 crc kubenswrapper[4761]: I1125 23:11:42.789825 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:42 crc kubenswrapper[4761]: I1125 23:11:42.791171 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:42 crc kubenswrapper[4761]: I1125 23:11:42.793627 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-7mzz2" Nov 25 23:11:42 crc kubenswrapper[4761]: I1125 23:11:42.803929 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:42 crc kubenswrapper[4761]: I1125 23:11:42.909007 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxw2d\" (UniqueName: \"kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d\") pod \"horizon-operator-index-5q2m5\" (UID: \"022820ad-f4ff-4c74-824a-40311fb99c66\") " pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.010947 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxw2d\" (UniqueName: \"kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d\") pod \"horizon-operator-index-5q2m5\" (UID: \"022820ad-f4ff-4c74-824a-40311fb99c66\") " pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.030632 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxw2d\" (UniqueName: \"kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d\") pod \"horizon-operator-index-5q2m5\" (UID: \"022820ad-f4ff-4c74-824a-40311fb99c66\") " pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.125779 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.261563 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.267876 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.389293 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:43 crc kubenswrapper[4761]: W1125 23:11:43.397011 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod022820ad_f4ff_4c74_824a_40311fb99c66.slice/crio-9be99f1fd1c4b91431fa747ebdae3a7b11175ea717eee4443bc120b7a938be98 WatchSource:0}: Error finding container 9be99f1fd1c4b91431fa747ebdae3a7b11175ea717eee4443bc120b7a938be98: Status 404 returned error can't find the container with id 9be99f1fd1c4b91431fa747ebdae3a7b11175ea717eee4443bc120b7a938be98 Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.415828 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrpx\" (UniqueName: \"kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx\") pod \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.416013 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqd86\" (UniqueName: \"kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86\") pod \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.416069 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts\") pod \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\" (UID: \"cda8b5ee-b6f4-4eb2-a095-f9647da330f6\") " Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.416103 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts\") pod \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\" (UID: \"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f\") " Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.417215 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" (UID: "32d39ec9-e2eb-41a4-83d8-f0f84fb3681f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.417555 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cda8b5ee-b6f4-4eb2-a095-f9647da330f6" (UID: "cda8b5ee-b6f4-4eb2-a095-f9647da330f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.420050 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx" (OuterVolumeSpecName: "kube-api-access-zxrpx") pod "cda8b5ee-b6f4-4eb2-a095-f9647da330f6" (UID: "cda8b5ee-b6f4-4eb2-a095-f9647da330f6"). InnerVolumeSpecName "kube-api-access-zxrpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.420820 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86" (OuterVolumeSpecName: "kube-api-access-zqd86") pod "32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" (UID: "32d39ec9-e2eb-41a4-83d8-f0f84fb3681f"). InnerVolumeSpecName "kube-api-access-zqd86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.517898 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrpx\" (UniqueName: \"kubernetes.io/projected/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-kube-api-access-zxrpx\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.518286 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqd86\" (UniqueName: \"kubernetes.io/projected/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-kube-api-access-zqd86\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.518301 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cda8b5ee-b6f4-4eb2-a095-f9647da330f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.518312 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.869372 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" event={"ID":"cda8b5ee-b6f4-4eb2-a095-f9647da330f6","Type":"ContainerDied","Data":"e515ccb3c4f8e11fb787cea8cb9c7ec7af93cb3e2b112bd613be491885ff4697"} Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.869436 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.869449 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e515ccb3c4f8e11fb787cea8cb9c7ec7af93cb3e2b112bd613be491885ff4697" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.871308 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hdrz6" event={"ID":"32d39ec9-e2eb-41a4-83d8-f0f84fb3681f","Type":"ContainerDied","Data":"0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9"} Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.871348 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hdrz6" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.871367 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d1eac11aac086bcc0caa55130425c6ba3232371bb28873556cfc9e4385456d9" Nov 25 23:11:43 crc kubenswrapper[4761]: I1125 23:11:43.873013 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-5q2m5" event={"ID":"022820ad-f4ff-4c74-824a-40311fb99c66","Type":"ContainerStarted","Data":"9be99f1fd1c4b91431fa747ebdae3a7b11175ea717eee4443bc120b7a938be98"} Nov 25 23:11:44 crc kubenswrapper[4761]: I1125 23:11:44.883336 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-5q2m5" event={"ID":"022820ad-f4ff-4c74-824a-40311fb99c66","Type":"ContainerStarted","Data":"689827138fbfabadf59a0975079bf782e4c3bf8afc8f40b4e2097c6ffea671d7"} Nov 25 23:11:44 crc kubenswrapper[4761]: I1125 23:11:44.913991 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-5q2m5" podStartSLOduration=1.931550868 podStartE2EDuration="2.913960986s" podCreationTimestamp="2025-11-25 23:11:42 +0000 UTC" firstStartedPulling="2025-11-25 23:11:43.399006173 +0000 UTC m=+899.132352008" lastFinishedPulling="2025-11-25 23:11:44.381416251 +0000 UTC m=+900.114762126" observedRunningTime="2025-11-25 23:11:44.905324777 +0000 UTC m=+900.638670692" watchObservedRunningTime="2025-11-25 23:11:44.913960986 +0000 UTC m=+900.647306861" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.666490 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6kdq"] Nov 25 23:11:45 crc kubenswrapper[4761]: E1125 23:11:45.667210 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" containerName="mariadb-database-create" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.667240 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" containerName="mariadb-database-create" Nov 25 23:11:45 crc kubenswrapper[4761]: E1125 23:11:45.667277 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cda8b5ee-b6f4-4eb2-a095-f9647da330f6" containerName="mariadb-account-create-update" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.667289 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="cda8b5ee-b6f4-4eb2-a095-f9647da330f6" containerName="mariadb-account-create-update" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.667470 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="cda8b5ee-b6f4-4eb2-a095-f9647da330f6" containerName="mariadb-account-create-update" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.667511 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" containerName="mariadb-database-create" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.668202 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.671438 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-txgjr" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.671607 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.672106 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.676890 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6kdq"] Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.683178 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.752873 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.753189 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgdx5\" (UniqueName: \"kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.854475 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.855544 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgdx5\" (UniqueName: \"kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.861077 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.873548 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgdx5\" (UniqueName: \"kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5\") pod \"keystone-db-sync-j6kdq\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:45 crc kubenswrapper[4761]: I1125 23:11:45.987746 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.174297 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-6z88v"] Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.175300 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.179490 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-mfh54" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.181955 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-6z88v"] Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.261221 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5f7n\" (UniqueName: \"kubernetes.io/projected/09f2d102-2c78-4217-890e-39872745a907-kube-api-access-r5f7n\") pod \"swift-operator-index-6z88v\" (UID: \"09f2d102-2c78-4217-890e-39872745a907\") " pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.362808 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5f7n\" (UniqueName: \"kubernetes.io/projected/09f2d102-2c78-4217-890e-39872745a907-kube-api-access-r5f7n\") pod \"swift-operator-index-6z88v\" (UID: \"09f2d102-2c78-4217-890e-39872745a907\") " pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.383147 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5f7n\" (UniqueName: \"kubernetes.io/projected/09f2d102-2c78-4217-890e-39872745a907-kube-api-access-r5f7n\") pod \"swift-operator-index-6z88v\" (UID: \"09f2d102-2c78-4217-890e-39872745a907\") " pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.430339 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6kdq"] Nov 25 23:11:46 crc kubenswrapper[4761]: W1125 23:11:46.437958 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30de3be7_530b_4f0b_910b_eb1060028d38.slice/crio-b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c WatchSource:0}: Error finding container b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c: Status 404 returned error can't find the container with id b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.503791 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.901344 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" event={"ID":"30de3be7-530b-4f0b-910b-eb1060028d38","Type":"ContainerStarted","Data":"b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c"} Nov 25 23:11:46 crc kubenswrapper[4761]: I1125 23:11:46.935832 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-6z88v"] Nov 25 23:11:46 crc kubenswrapper[4761]: W1125 23:11:46.946401 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09f2d102_2c78_4217_890e_39872745a907.slice/crio-e4121ec6345afc4a61b9d1923e5748e26bb55ecfe66ae1bb6d3cde24bdfc6ef2 WatchSource:0}: Error finding container e4121ec6345afc4a61b9d1923e5748e26bb55ecfe66ae1bb6d3cde24bdfc6ef2: Status 404 returned error can't find the container with id e4121ec6345afc4a61b9d1923e5748e26bb55ecfe66ae1bb6d3cde24bdfc6ef2 Nov 25 23:11:47 crc kubenswrapper[4761]: I1125 23:11:47.908148 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-6z88v" event={"ID":"09f2d102-2c78-4217-890e-39872745a907","Type":"ContainerStarted","Data":"e4121ec6345afc4a61b9d1923e5748e26bb55ecfe66ae1bb6d3cde24bdfc6ef2"} Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.378444 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.378845 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-5q2m5" podUID="022820ad-f4ff-4c74-824a-40311fb99c66" containerName="registry-server" containerID="cri-o://689827138fbfabadf59a0975079bf782e4c3bf8afc8f40b4e2097c6ffea671d7" gracePeriod=2 Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.921861 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-6z88v" event={"ID":"09f2d102-2c78-4217-890e-39872745a907","Type":"ContainerStarted","Data":"647030ed408239b63a136ec0f90ac67667935a91e838706a71903f19cd78fe07"} Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.934592 4761 generic.go:334] "Generic (PLEG): container finished" podID="022820ad-f4ff-4c74-824a-40311fb99c66" containerID="689827138fbfabadf59a0975079bf782e4c3bf8afc8f40b4e2097c6ffea671d7" exitCode=0 Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.934641 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-5q2m5" event={"ID":"022820ad-f4ff-4c74-824a-40311fb99c66","Type":"ContainerDied","Data":"689827138fbfabadf59a0975079bf782e4c3bf8afc8f40b4e2097c6ffea671d7"} Nov 25 23:11:48 crc kubenswrapper[4761]: I1125 23:11:48.939023 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-6z88v" podStartSLOduration=2.128019727 podStartE2EDuration="2.939001199s" podCreationTimestamp="2025-11-25 23:11:46 +0000 UTC" firstStartedPulling="2025-11-25 23:11:46.95005665 +0000 UTC m=+902.683402525" lastFinishedPulling="2025-11-25 23:11:47.761038162 +0000 UTC m=+903.494383997" observedRunningTime="2025-11-25 23:11:48.93565238 +0000 UTC m=+904.668998235" watchObservedRunningTime="2025-11-25 23:11:48.939001199 +0000 UTC m=+904.672347044" Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.176746 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-pkh4w"] Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.177664 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.189153 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-pkh4w"] Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.311769 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhd8x\" (UniqueName: \"kubernetes.io/projected/ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad-kube-api-access-qhd8x\") pod \"horizon-operator-index-pkh4w\" (UID: \"ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad\") " pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.414001 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhd8x\" (UniqueName: \"kubernetes.io/projected/ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad-kube-api-access-qhd8x\") pod \"horizon-operator-index-pkh4w\" (UID: \"ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad\") " pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.439896 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhd8x\" (UniqueName: \"kubernetes.io/projected/ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad-kube-api-access-qhd8x\") pod \"horizon-operator-index-pkh4w\" (UID: \"ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad\") " pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:49 crc kubenswrapper[4761]: I1125 23:11:49.502398 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:51 crc kubenswrapper[4761]: I1125 23:11:51.128249 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:11:51 crc kubenswrapper[4761]: I1125 23:11:51.128747 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.126810 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.199711 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.386760 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxw2d\" (UniqueName: \"kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d\") pod \"022820ad-f4ff-4c74-824a-40311fb99c66\" (UID: \"022820ad-f4ff-4c74-824a-40311fb99c66\") " Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.391989 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d" (OuterVolumeSpecName: "kube-api-access-lxw2d") pod "022820ad-f4ff-4c74-824a-40311fb99c66" (UID: "022820ad-f4ff-4c74-824a-40311fb99c66"). InnerVolumeSpecName "kube-api-access-lxw2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.488869 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxw2d\" (UniqueName: \"kubernetes.io/projected/022820ad-f4ff-4c74-824a-40311fb99c66-kube-api-access-lxw2d\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.606521 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-pkh4w"] Nov 25 23:11:53 crc kubenswrapper[4761]: W1125 23:11:53.616900 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba0b63b3_4cf1_42f4_8daf_1b10aca0e4ad.slice/crio-19e6b5f19410cd250872af3025259fef765da5f5f62a73270a2b0b234f33ef84 WatchSource:0}: Error finding container 19e6b5f19410cd250872af3025259fef765da5f5f62a73270a2b0b234f33ef84: Status 404 returned error can't find the container with id 19e6b5f19410cd250872af3025259fef765da5f5f62a73270a2b0b234f33ef84 Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.986375 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-5q2m5" event={"ID":"022820ad-f4ff-4c74-824a-40311fb99c66","Type":"ContainerDied","Data":"9be99f1fd1c4b91431fa747ebdae3a7b11175ea717eee4443bc120b7a938be98"} Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.986458 4761 scope.go:117] "RemoveContainer" containerID="689827138fbfabadf59a0975079bf782e4c3bf8afc8f40b4e2097c6ffea671d7" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.986391 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-5q2m5" Nov 25 23:11:53 crc kubenswrapper[4761]: I1125 23:11:53.994436 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" event={"ID":"30de3be7-530b-4f0b-910b-eb1060028d38","Type":"ContainerStarted","Data":"f038d6766a854ff30e58dd0bccbecc637071ad169e27825d17deb52063b02949"} Nov 25 23:11:54 crc kubenswrapper[4761]: I1125 23:11:54.004612 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-pkh4w" event={"ID":"ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad","Type":"ContainerStarted","Data":"19e6b5f19410cd250872af3025259fef765da5f5f62a73270a2b0b234f33ef84"} Nov 25 23:11:54 crc kubenswrapper[4761]: I1125 23:11:54.031609 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" podStartSLOduration=2.249331281 podStartE2EDuration="9.030767255s" podCreationTimestamp="2025-11-25 23:11:45 +0000 UTC" firstStartedPulling="2025-11-25 23:11:46.440850667 +0000 UTC m=+902.174196502" lastFinishedPulling="2025-11-25 23:11:53.222286601 +0000 UTC m=+908.955632476" observedRunningTime="2025-11-25 23:11:54.02193159 +0000 UTC m=+909.755277505" watchObservedRunningTime="2025-11-25 23:11:54.030767255 +0000 UTC m=+909.764113120" Nov 25 23:11:54 crc kubenswrapper[4761]: I1125 23:11:54.050208 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:54 crc kubenswrapper[4761]: I1125 23:11:54.055012 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-5q2m5"] Nov 25 23:11:55 crc kubenswrapper[4761]: I1125 23:11:55.026679 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="022820ad-f4ff-4c74-824a-40311fb99c66" path="/var/lib/kubelet/pods/022820ad-f4ff-4c74-824a-40311fb99c66/volumes" Nov 25 23:11:55 crc kubenswrapper[4761]: I1125 23:11:55.027928 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-pkh4w" event={"ID":"ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad","Type":"ContainerStarted","Data":"69e04e7760dc839fa690fb1a38d4391cbf60d3527b0be3dcf46b5a59d08b64cc"} Nov 25 23:11:55 crc kubenswrapper[4761]: I1125 23:11:55.078623 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-pkh4w" podStartSLOduration=5.647316468 podStartE2EDuration="6.078591785s" podCreationTimestamp="2025-11-25 23:11:49 +0000 UTC" firstStartedPulling="2025-11-25 23:11:53.621401312 +0000 UTC m=+909.354747157" lastFinishedPulling="2025-11-25 23:11:54.052676629 +0000 UTC m=+909.786022474" observedRunningTime="2025-11-25 23:11:55.070521332 +0000 UTC m=+910.803867237" watchObservedRunningTime="2025-11-25 23:11:55.078591785 +0000 UTC m=+910.811937660" Nov 25 23:11:56 crc kubenswrapper[4761]: I1125 23:11:56.504831 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:56 crc kubenswrapper[4761]: I1125 23:11:56.505210 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:56 crc kubenswrapper[4761]: I1125 23:11:56.538402 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:57 crc kubenswrapper[4761]: I1125 23:11:57.071312 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-6z88v" Nov 25 23:11:58 crc kubenswrapper[4761]: I1125 23:11:58.042094 4761 generic.go:334] "Generic (PLEG): container finished" podID="30de3be7-530b-4f0b-910b-eb1060028d38" containerID="f038d6766a854ff30e58dd0bccbecc637071ad169e27825d17deb52063b02949" exitCode=0 Nov 25 23:11:58 crc kubenswrapper[4761]: I1125 23:11:58.042158 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" event={"ID":"30de3be7-530b-4f0b-910b-eb1060028d38","Type":"ContainerDied","Data":"f038d6766a854ff30e58dd0bccbecc637071ad169e27825d17deb52063b02949"} Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.447357 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.502999 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.503072 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.545680 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.598020 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgdx5\" (UniqueName: \"kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5\") pod \"30de3be7-530b-4f0b-910b-eb1060028d38\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.598166 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data\") pod \"30de3be7-530b-4f0b-910b-eb1060028d38\" (UID: \"30de3be7-530b-4f0b-910b-eb1060028d38\") " Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.607954 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5" (OuterVolumeSpecName: "kube-api-access-jgdx5") pod "30de3be7-530b-4f0b-910b-eb1060028d38" (UID: "30de3be7-530b-4f0b-910b-eb1060028d38"). InnerVolumeSpecName "kube-api-access-jgdx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.644338 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data" (OuterVolumeSpecName: "config-data") pod "30de3be7-530b-4f0b-910b-eb1060028d38" (UID: "30de3be7-530b-4f0b-910b-eb1060028d38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.699462 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30de3be7-530b-4f0b-910b-eb1060028d38-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:11:59 crc kubenswrapper[4761]: I1125 23:11:59.699505 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgdx5\" (UniqueName: \"kubernetes.io/projected/30de3be7-530b-4f0b-910b-eb1060028d38-kube-api-access-jgdx5\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.061033 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.061025 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6kdq" event={"ID":"30de3be7-530b-4f0b-910b-eb1060028d38","Type":"ContainerDied","Data":"b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c"} Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.061113 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b734073c5099f61d6913964f1b4a48d36d768917d392451f332e792a086d798c" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.121038 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-pkh4w" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.285947 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-prggk"] Nov 25 23:12:00 crc kubenswrapper[4761]: E1125 23:12:00.286474 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="022820ad-f4ff-4c74-824a-40311fb99c66" containerName="registry-server" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.286564 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="022820ad-f4ff-4c74-824a-40311fb99c66" containerName="registry-server" Nov 25 23:12:00 crc kubenswrapper[4761]: E1125 23:12:00.286659 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30de3be7-530b-4f0b-910b-eb1060028d38" containerName="keystone-db-sync" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.286764 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="30de3be7-530b-4f0b-910b-eb1060028d38" containerName="keystone-db-sync" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.287006 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="30de3be7-530b-4f0b-910b-eb1060028d38" containerName="keystone-db-sync" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.287096 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="022820ad-f4ff-4c74-824a-40311fb99c66" containerName="registry-server" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.287655 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.290325 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.290782 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-txgjr" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.290939 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.291158 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.291325 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.306961 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-prggk"] Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.310273 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.310473 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz2qm\" (UniqueName: \"kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.310589 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.310711 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.310804 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.411502 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.411747 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.411885 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.412521 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz2qm\" (UniqueName: \"kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.412611 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.414993 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.424767 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.425187 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.425621 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.435801 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz2qm\" (UniqueName: \"kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm\") pod \"keystone-bootstrap-prggk\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:00 crc kubenswrapper[4761]: I1125 23:12:00.603176 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:01 crc kubenswrapper[4761]: I1125 23:12:01.110784 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-prggk"] Nov 25 23:12:02 crc kubenswrapper[4761]: I1125 23:12:02.080775 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-prggk" event={"ID":"a4fd046a-5f86-4226-a001-c6a166b2fbad","Type":"ContainerStarted","Data":"5820a2faf3311a75bd9ee55461a4b38580800d93ff693c1ca9778f76144a3284"} Nov 25 23:12:03 crc kubenswrapper[4761]: I1125 23:12:03.089449 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-prggk" event={"ID":"a4fd046a-5f86-4226-a001-c6a166b2fbad","Type":"ContainerStarted","Data":"774285aecbee5bfcad39e7ed1cbfb99bbb4ac2059dfae882e8e6ee368d0d4481"} Nov 25 23:12:03 crc kubenswrapper[4761]: I1125 23:12:03.122417 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-prggk" podStartSLOduration=3.122393474 podStartE2EDuration="3.122393474s" podCreationTimestamp="2025-11-25 23:12:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:12:03.117746211 +0000 UTC m=+918.851092136" watchObservedRunningTime="2025-11-25 23:12:03.122393474 +0000 UTC m=+918.855739349" Nov 25 23:12:06 crc kubenswrapper[4761]: I1125 23:12:06.128066 4761 generic.go:334] "Generic (PLEG): container finished" podID="a4fd046a-5f86-4226-a001-c6a166b2fbad" containerID="774285aecbee5bfcad39e7ed1cbfb99bbb4ac2059dfae882e8e6ee368d0d4481" exitCode=0 Nov 25 23:12:06 crc kubenswrapper[4761]: I1125 23:12:06.128133 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-prggk" event={"ID":"a4fd046a-5f86-4226-a001-c6a166b2fbad","Type":"ContainerDied","Data":"774285aecbee5bfcad39e7ed1cbfb99bbb4ac2059dfae882e8e6ee368d0d4481"} Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.535889 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.722681 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys\") pod \"a4fd046a-5f86-4226-a001-c6a166b2fbad\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.722835 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts\") pod \"a4fd046a-5f86-4226-a001-c6a166b2fbad\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.722968 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys\") pod \"a4fd046a-5f86-4226-a001-c6a166b2fbad\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.723004 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data\") pod \"a4fd046a-5f86-4226-a001-c6a166b2fbad\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.723045 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz2qm\" (UniqueName: \"kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm\") pod \"a4fd046a-5f86-4226-a001-c6a166b2fbad\" (UID: \"a4fd046a-5f86-4226-a001-c6a166b2fbad\") " Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.730966 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm" (OuterVolumeSpecName: "kube-api-access-zz2qm") pod "a4fd046a-5f86-4226-a001-c6a166b2fbad" (UID: "a4fd046a-5f86-4226-a001-c6a166b2fbad"). InnerVolumeSpecName "kube-api-access-zz2qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.731249 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a4fd046a-5f86-4226-a001-c6a166b2fbad" (UID: "a4fd046a-5f86-4226-a001-c6a166b2fbad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.742875 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts" (OuterVolumeSpecName: "scripts") pod "a4fd046a-5f86-4226-a001-c6a166b2fbad" (UID: "a4fd046a-5f86-4226-a001-c6a166b2fbad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.743896 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a4fd046a-5f86-4226-a001-c6a166b2fbad" (UID: "a4fd046a-5f86-4226-a001-c6a166b2fbad"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.758256 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data" (OuterVolumeSpecName: "config-data") pod "a4fd046a-5f86-4226-a001-c6a166b2fbad" (UID: "a4fd046a-5f86-4226-a001-c6a166b2fbad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.825598 4761 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.825645 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.825663 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz2qm\" (UniqueName: \"kubernetes.io/projected/a4fd046a-5f86-4226-a001-c6a166b2fbad-kube-api-access-zz2qm\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.825679 4761 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:07 crc kubenswrapper[4761]: I1125 23:12:07.825726 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4fd046a-5f86-4226-a001-c6a166b2fbad-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.149325 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-prggk" event={"ID":"a4fd046a-5f86-4226-a001-c6a166b2fbad","Type":"ContainerDied","Data":"5820a2faf3311a75bd9ee55461a4b38580800d93ff693c1ca9778f76144a3284"} Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.149369 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5820a2faf3311a75bd9ee55461a4b38580800d93ff693c1ca9778f76144a3284" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.149391 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-prggk" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.257695 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cdfb754f5-ddx65"] Nov 25 23:12:08 crc kubenswrapper[4761]: E1125 23:12:08.258098 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fd046a-5f86-4226-a001-c6a166b2fbad" containerName="keystone-bootstrap" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.258119 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fd046a-5f86-4226-a001-c6a166b2fbad" containerName="keystone-bootstrap" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.258445 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fd046a-5f86-4226-a001-c6a166b2fbad" containerName="keystone-bootstrap" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.259208 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.262296 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.262952 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.263239 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.264410 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-txgjr" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.280890 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cdfb754f5-ddx65"] Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.434664 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-fernet-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.434742 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-credential-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.434819 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p79vs\" (UniqueName: \"kubernetes.io/projected/d8846a3a-456b-480c-b911-6e205ff677ab-kube-api-access-p79vs\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.434857 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-config-data\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.434896 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-scripts\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.536184 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-scripts\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.537887 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-fernet-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.538159 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-credential-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.538389 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p79vs\" (UniqueName: \"kubernetes.io/projected/d8846a3a-456b-480c-b911-6e205ff677ab-kube-api-access-p79vs\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.538647 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-config-data\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.541919 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-scripts\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.544252 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-credential-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.544444 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-fernet-keys\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.544474 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8846a3a-456b-480c-b911-6e205ff677ab-config-data\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.564454 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p79vs\" (UniqueName: \"kubernetes.io/projected/d8846a3a-456b-480c-b911-6e205ff677ab-kube-api-access-p79vs\") pod \"keystone-cdfb754f5-ddx65\" (UID: \"d8846a3a-456b-480c-b911-6e205ff677ab\") " pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.588198 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.811862 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2"] Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.813498 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.817852 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.824537 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2"] Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.890306 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cdfb754f5-ddx65"] Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.945455 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.945551 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fmvl\" (UniqueName: \"kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:08 crc kubenswrapper[4761]: I1125 23:12:08.945600 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.047060 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fmvl\" (UniqueName: \"kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.047156 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.047211 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.047833 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.047978 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.063082 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fmvl\" (UniqueName: \"kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.137347 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.177241 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" event={"ID":"d8846a3a-456b-480c-b911-6e205ff677ab","Type":"ContainerStarted","Data":"abc76a3ce20dcff8cbd1991f9a5e33d081fadc6f4276384e4c6493c39830396b"} Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.177299 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" event={"ID":"d8846a3a-456b-480c-b911-6e205ff677ab","Type":"ContainerStarted","Data":"76471418ee834bf4032939424471dd132f9132d3f0d7060dc696246b54fc675f"} Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.177847 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.205477 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" podStartSLOduration=1.205451548 podStartE2EDuration="1.205451548s" podCreationTimestamp="2025-11-25 23:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:12:09.193989557 +0000 UTC m=+924.927335402" watchObservedRunningTime="2025-11-25 23:12:09.205451548 +0000 UTC m=+924.938797393" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.604888 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2"] Nov 25 23:12:09 crc kubenswrapper[4761]: W1125 23:12:09.617675 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf932b8b_53ef_4e97_ac65_ec51f2177155.slice/crio-b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3 WatchSource:0}: Error finding container b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3: Status 404 returned error can't find the container with id b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3 Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.832991 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4"] Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.835052 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.854779 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4"] Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.961598 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.961666 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:09 crc kubenswrapper[4761]: I1125 23:12:09.961982 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmp4m\" (UniqueName: \"kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.063285 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmp4m\" (UniqueName: \"kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.063400 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.063483 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.064233 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.064257 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.088310 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmp4m\" (UniqueName: \"kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.152644 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.184184 4761 generic.go:334] "Generic (PLEG): container finished" podID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerID="c5d2a44c40e83f158bf6fffec42f302bd22c06d8ad3787e463c14bf9079f1527" exitCode=0 Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.185075 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" event={"ID":"cf932b8b-53ef-4e97-ac65-ec51f2177155","Type":"ContainerDied","Data":"c5d2a44c40e83f158bf6fffec42f302bd22c06d8ad3787e463c14bf9079f1527"} Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.185108 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" event={"ID":"cf932b8b-53ef-4e97-ac65-ec51f2177155","Type":"ContainerStarted","Data":"b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3"} Nov 25 23:12:10 crc kubenswrapper[4761]: I1125 23:12:10.586208 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4"] Nov 25 23:12:11 crc kubenswrapper[4761]: I1125 23:12:11.194814 4761 generic.go:334] "Generic (PLEG): container finished" podID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerID="01905b62a572810c6d8228dcf22270bce0a3772912b4e95eecff96fc694cb026" exitCode=0 Nov 25 23:12:11 crc kubenswrapper[4761]: I1125 23:12:11.194891 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" event={"ID":"fe7636ab-8296-4e62-bba7-61ccb3507e9f","Type":"ContainerDied","Data":"01905b62a572810c6d8228dcf22270bce0a3772912b4e95eecff96fc694cb026"} Nov 25 23:12:11 crc kubenswrapper[4761]: I1125 23:12:11.195255 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" event={"ID":"fe7636ab-8296-4e62-bba7-61ccb3507e9f","Type":"ContainerStarted","Data":"d15d268509317b90be90c6a4a42261332adef226369f94ee3d6025889bbcf77f"} Nov 25 23:12:11 crc kubenswrapper[4761]: I1125 23:12:11.197618 4761 generic.go:334] "Generic (PLEG): container finished" podID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerID="ad7f9d57bf0bb6c1f8bc1582cf40649cade644e75629dcb4cedd02b1e1a9a965" exitCode=0 Nov 25 23:12:11 crc kubenswrapper[4761]: I1125 23:12:11.197671 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" event={"ID":"cf932b8b-53ef-4e97-ac65-ec51f2177155","Type":"ContainerDied","Data":"ad7f9d57bf0bb6c1f8bc1582cf40649cade644e75629dcb4cedd02b1e1a9a965"} Nov 25 23:12:12 crc kubenswrapper[4761]: I1125 23:12:12.210099 4761 generic.go:334] "Generic (PLEG): container finished" podID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerID="b2d08eab4769e15805ec20cb599a2c41e82cfb67488633715222f1c4315079e5" exitCode=0 Nov 25 23:12:12 crc kubenswrapper[4761]: I1125 23:12:12.210179 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" event={"ID":"cf932b8b-53ef-4e97-ac65-ec51f2177155","Type":"ContainerDied","Data":"b2d08eab4769e15805ec20cb599a2c41e82cfb67488633715222f1c4315079e5"} Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.219829 4761 generic.go:334] "Generic (PLEG): container finished" podID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerID="b0420f9a99da267d7d82c9141623060bfff933eb69d4940fc186fe280c2c6f2f" exitCode=0 Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.219910 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" event={"ID":"fe7636ab-8296-4e62-bba7-61ccb3507e9f","Type":"ContainerDied","Data":"b0420f9a99da267d7d82c9141623060bfff933eb69d4940fc186fe280c2c6f2f"} Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.674337 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.821376 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util\") pod \"cf932b8b-53ef-4e97-ac65-ec51f2177155\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.821490 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fmvl\" (UniqueName: \"kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl\") pod \"cf932b8b-53ef-4e97-ac65-ec51f2177155\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.821626 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle\") pod \"cf932b8b-53ef-4e97-ac65-ec51f2177155\" (UID: \"cf932b8b-53ef-4e97-ac65-ec51f2177155\") " Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.822671 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle" (OuterVolumeSpecName: "bundle") pod "cf932b8b-53ef-4e97-ac65-ec51f2177155" (UID: "cf932b8b-53ef-4e97-ac65-ec51f2177155"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.830365 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl" (OuterVolumeSpecName: "kube-api-access-9fmvl") pod "cf932b8b-53ef-4e97-ac65-ec51f2177155" (UID: "cf932b8b-53ef-4e97-ac65-ec51f2177155"). InnerVolumeSpecName "kube-api-access-9fmvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.847270 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util" (OuterVolumeSpecName: "util") pod "cf932b8b-53ef-4e97-ac65-ec51f2177155" (UID: "cf932b8b-53ef-4e97-ac65-ec51f2177155"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.924165 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.924211 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fmvl\" (UniqueName: \"kubernetes.io/projected/cf932b8b-53ef-4e97-ac65-ec51f2177155-kube-api-access-9fmvl\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:13 crc kubenswrapper[4761]: I1125 23:12:13.924229 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf932b8b-53ef-4e97-ac65-ec51f2177155-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:14 crc kubenswrapper[4761]: I1125 23:12:14.231727 4761 generic.go:334] "Generic (PLEG): container finished" podID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerID="12c20ec42e6b487659d4a3573914c0834144a751a78ea65e0017de9ba5a16e62" exitCode=0 Nov 25 23:12:14 crc kubenswrapper[4761]: I1125 23:12:14.231786 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" event={"ID":"fe7636ab-8296-4e62-bba7-61ccb3507e9f","Type":"ContainerDied","Data":"12c20ec42e6b487659d4a3573914c0834144a751a78ea65e0017de9ba5a16e62"} Nov 25 23:12:14 crc kubenswrapper[4761]: I1125 23:12:14.235193 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" event={"ID":"cf932b8b-53ef-4e97-ac65-ec51f2177155","Type":"ContainerDied","Data":"b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3"} Nov 25 23:12:14 crc kubenswrapper[4761]: I1125 23:12:14.235222 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b15b89355fc5d4292c5ac2b6df87cfa28fa702c40e9bdffa293e15ca4c526fd3" Nov 25 23:12:14 crc kubenswrapper[4761]: I1125 23:12:14.235308 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.570193 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.648069 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmp4m\" (UniqueName: \"kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m\") pod \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.648177 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util\") pod \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.648284 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle\") pod \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\" (UID: \"fe7636ab-8296-4e62-bba7-61ccb3507e9f\") " Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.649882 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle" (OuterVolumeSpecName: "bundle") pod "fe7636ab-8296-4e62-bba7-61ccb3507e9f" (UID: "fe7636ab-8296-4e62-bba7-61ccb3507e9f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.652117 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m" (OuterVolumeSpecName: "kube-api-access-lmp4m") pod "fe7636ab-8296-4e62-bba7-61ccb3507e9f" (UID: "fe7636ab-8296-4e62-bba7-61ccb3507e9f"). InnerVolumeSpecName "kube-api-access-lmp4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.664088 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util" (OuterVolumeSpecName: "util") pod "fe7636ab-8296-4e62-bba7-61ccb3507e9f" (UID: "fe7636ab-8296-4e62-bba7-61ccb3507e9f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.750204 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.750249 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmp4m\" (UniqueName: \"kubernetes.io/projected/fe7636ab-8296-4e62-bba7-61ccb3507e9f-kube-api-access-lmp4m\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:15 crc kubenswrapper[4761]: I1125 23:12:15.750263 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe7636ab-8296-4e62-bba7-61ccb3507e9f-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:16 crc kubenswrapper[4761]: I1125 23:12:16.249991 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" event={"ID":"fe7636ab-8296-4e62-bba7-61ccb3507e9f","Type":"ContainerDied","Data":"d15d268509317b90be90c6a4a42261332adef226369f94ee3d6025889bbcf77f"} Nov 25 23:12:16 crc kubenswrapper[4761]: I1125 23:12:16.250032 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15d268509317b90be90c6a4a42261332adef226369f94ee3d6025889bbcf77f" Nov 25 23:12:16 crc kubenswrapper[4761]: I1125 23:12:16.250053 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4" Nov 25 23:12:21 crc kubenswrapper[4761]: I1125 23:12:21.128213 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:12:21 crc kubenswrapper[4761]: I1125 23:12:21.128635 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.892262 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc"] Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893180 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="pull" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893196 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="pull" Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893217 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="pull" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893225 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="pull" Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893240 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893249 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893260 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="util" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893268 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="util" Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893286 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="util" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893293 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="util" Nov 25 23:12:33 crc kubenswrapper[4761]: E1125 23:12:33.893308 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893316 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893438 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7636ab-8296-4e62-bba7-61ccb3507e9f" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893457 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf932b8b-53ef-4e97-ac65-ec51f2177155" containerName="extract" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.893930 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.896820 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.897258 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-t4jrq" Nov 25 23:12:33 crc kubenswrapper[4761]: I1125 23:12:33.908733 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc"] Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.047228 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-apiservice-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.047581 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-webhook-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.047613 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb4cg\" (UniqueName: \"kubernetes.io/projected/8aba4458-8023-4b9f-8504-6fab49f5250c-kube-api-access-pb4cg\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.148691 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-apiservice-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.148812 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-webhook-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.148836 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb4cg\" (UniqueName: \"kubernetes.io/projected/8aba4458-8023-4b9f-8504-6fab49f5250c-kube-api-access-pb4cg\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.154328 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-webhook-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.155168 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8aba4458-8023-4b9f-8504-6fab49f5250c-apiservice-cert\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.166991 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb4cg\" (UniqueName: \"kubernetes.io/projected/8aba4458-8023-4b9f-8504-6fab49f5250c-kube-api-access-pb4cg\") pod \"swift-operator-controller-manager-6c6d6c68d5-v5cfc\" (UID: \"8aba4458-8023-4b9f-8504-6fab49f5250c\") " pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.214538 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.503494 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn"] Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.504653 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.507222 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-g4w2q" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.509060 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.515628 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn"] Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.645996 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc"] Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.655820 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-apiservice-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.656218 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkxsj\" (UniqueName: \"kubernetes.io/projected/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-kube-api-access-wkxsj\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.656275 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-webhook-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.657145 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.758026 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-apiservice-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.758291 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkxsj\" (UniqueName: \"kubernetes.io/projected/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-kube-api-access-wkxsj\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.758382 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-webhook-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.762531 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-apiservice-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.763158 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-webhook-cert\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.785344 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkxsj\" (UniqueName: \"kubernetes.io/projected/bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b-kube-api-access-wkxsj\") pod \"horizon-operator-controller-manager-7f54679d79-6ldpn\" (UID: \"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b\") " pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:34 crc kubenswrapper[4761]: I1125 23:12:34.819904 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:35 crc kubenswrapper[4761]: I1125 23:12:35.230801 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn"] Nov 25 23:12:35 crc kubenswrapper[4761]: I1125 23:12:35.407319 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerStarted","Data":"4d96d5129331303071fdfcdd6c13111a690f6032d79fbe31771814a2c1eb238a"} Nov 25 23:12:35 crc kubenswrapper[4761]: I1125 23:12:35.408468 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerStarted","Data":"184aad5c154232ed7800aed304a960286dd6a0de044dfa4a8c0379d4096d84ed"} Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.427127 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerStarted","Data":"80ddb3e8634b4d68c9f701a1c14b8be0415779a56d956cee8b895f3962cdda58"} Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.427448 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.430114 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerStarted","Data":"3d10673a484790e525cda35f8eb9fc95ab6f6dbc67bbaa621dbbe35ec976a66b"} Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.430733 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.464103 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podStartSLOduration=1.7036898 podStartE2EDuration="3.464083947s" podCreationTimestamp="2025-11-25 23:12:34 +0000 UTC" firstStartedPulling="2025-11-25 23:12:35.239599815 +0000 UTC m=+950.972945690" lastFinishedPulling="2025-11-25 23:12:36.999994002 +0000 UTC m=+952.733339837" observedRunningTime="2025-11-25 23:12:37.461807637 +0000 UTC m=+953.195153492" watchObservedRunningTime="2025-11-25 23:12:37.464083947 +0000 UTC m=+953.197429792" Nov 25 23:12:37 crc kubenswrapper[4761]: I1125 23:12:37.524989 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podStartSLOduration=2.186184228 podStartE2EDuration="4.52496783s" podCreationTimestamp="2025-11-25 23:12:33 +0000 UTC" firstStartedPulling="2025-11-25 23:12:34.656747483 +0000 UTC m=+950.390093358" lastFinishedPulling="2025-11-25 23:12:36.995531135 +0000 UTC m=+952.728876960" observedRunningTime="2025-11-25 23:12:37.521945591 +0000 UTC m=+953.255291436" watchObservedRunningTime="2025-11-25 23:12:37.52496783 +0000 UTC m=+953.258313685" Nov 25 23:12:39 crc kubenswrapper[4761]: I1125 23:12:39.991923 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-cdfb754f5-ddx65" Nov 25 23:12:44 crc kubenswrapper[4761]: I1125 23:12:44.220803 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:12:44 crc kubenswrapper[4761]: I1125 23:12:44.823531 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.634026 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.638074 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.643661 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.643807 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.643966 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.644014 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-72m9f" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.655253 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.764868 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.764937 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-lock\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.765219 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.765279 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-cache\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.765309 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8brm5\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-kube-api-access-8brm5\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.866911 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-cache\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.866957 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8brm5\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-kube-api-access-8brm5\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.866982 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: E1125 23:12:47.867290 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:47 crc kubenswrapper[4761]: E1125 23:12:47.867332 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.867342 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-lock\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.867387 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-cache\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: E1125 23:12:47.867398 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift podName:ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad nodeName:}" failed. No retries permitted until 2025-11-25 23:12:48.367375174 +0000 UTC m=+964.100721019 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift") pod "swift-storage-0" (UID: "ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad") : configmap "swift-ring-files" not found Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.867713 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-lock\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.867832 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.868095 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.898986 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8brm5\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-kube-api-access-8brm5\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:47 crc kubenswrapper[4761]: I1125 23:12:47.904788 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.260199 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-9mc86"] Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.261009 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.264774 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.269859 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.270688 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.279415 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-9mc86"] Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374519 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374576 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374602 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fz6w\" (UniqueName: \"kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374670 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374730 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374768 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: E1125 23:12:48.374774 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:48 crc kubenswrapper[4761]: E1125 23:12:48.374810 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.374811 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: E1125 23:12:48.374869 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift podName:ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad nodeName:}" failed. No retries permitted until 2025-11-25 23:12:49.37485057 +0000 UTC m=+965.108196405 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift") pod "swift-storage-0" (UID: "ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad") : configmap "swift-ring-files" not found Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.475835 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.475888 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.475924 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.475958 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.475978 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.476009 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fz6w\" (UniqueName: \"kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.476629 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.476873 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.476923 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.491089 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.491209 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.496106 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fz6w\" (UniqueName: \"kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w\") pod \"swift-ring-rebalance-9mc86\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:48 crc kubenswrapper[4761]: I1125 23:12:48.577147 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.118551 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-9mc86"] Nov 25 23:12:49 crc kubenswrapper[4761]: W1125 23:12:49.126442 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb067022b_e70b_4750_9d89_f9e48ca02b04.slice/crio-5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf WatchSource:0}: Error finding container 5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf: Status 404 returned error can't find the container with id 5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.177261 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.178077 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.180178 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-2tj2k" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.195363 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.285884 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8tsz\" (UniqueName: \"kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz\") pod \"glance-operator-index-bbmjg\" (UID: \"327a4f2c-0d8d-4938-aa58-008bfa75bcca\") " pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.288392 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl"] Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.289736 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.307727 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl"] Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387119 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387196 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8tsz\" (UniqueName: \"kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz\") pod \"glance-operator-index-bbmjg\" (UID: \"327a4f2c-0d8d-4938-aa58-008bfa75bcca\") " pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387225 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e65981-a172-41eb-ac4d-8bbe43248f0a-config-data\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387282 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-run-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387342 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387380 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw9zl\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-kube-api-access-cw9zl\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.387576 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-log-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.387636 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.387688 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.387837 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift podName:ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad nodeName:}" failed. No retries permitted until 2025-11-25 23:12:51.387797751 +0000 UTC m=+967.121143656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift") pod "swift-storage-0" (UID: "ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad") : configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.420414 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8tsz\" (UniqueName: \"kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz\") pod \"glance-operator-index-bbmjg\" (UID: \"327a4f2c-0d8d-4938-aa58-008bfa75bcca\") " pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489034 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-log-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489097 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489128 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e65981-a172-41eb-ac4d-8bbe43248f0a-config-data\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489166 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-run-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489228 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw9zl\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-kube-api-access-cw9zl\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.489317 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.489349 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.489433 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift podName:c7e65981-a172-41eb-ac4d-8bbe43248f0a nodeName:}" failed. No retries permitted until 2025-11-25 23:12:49.989408977 +0000 UTC m=+965.722754812 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift") pod "swift-proxy-6bd58cfcf7-dxxvl" (UID: "c7e65981-a172-41eb-ac4d-8bbe43248f0a") : configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489575 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-log-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.489608 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7e65981-a172-41eb-ac4d-8bbe43248f0a-run-httpd\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.493710 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e65981-a172-41eb-ac4d-8bbe43248f0a-config-data\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.493930 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.516590 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw9zl\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-kube-api-access-cw9zl\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.528974 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" event={"ID":"b067022b-e70b-4750-9d89-f9e48ca02b04","Type":"ContainerStarted","Data":"5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf"} Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.767319 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:49 crc kubenswrapper[4761]: I1125 23:12:49.995421 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.995719 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.995751 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl: configmap "swift-ring-files" not found Nov 25 23:12:49 crc kubenswrapper[4761]: E1125 23:12:49.995829 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift podName:c7e65981-a172-41eb-ac4d-8bbe43248f0a nodeName:}" failed. No retries permitted until 2025-11-25 23:12:50.995808526 +0000 UTC m=+966.729154351 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift") pod "swift-proxy-6bd58cfcf7-dxxvl" (UID: "c7e65981-a172-41eb-ac4d-8bbe43248f0a") : configmap "swift-ring-files" not found Nov 25 23:12:50 crc kubenswrapper[4761]: I1125 23:12:50.536142 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bbmjg" event={"ID":"327a4f2c-0d8d-4938-aa58-008bfa75bcca","Type":"ContainerStarted","Data":"311988575ce39513eec3008af25a524c1ff36a30913b661dc51f7fa9072122c8"} Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.011090 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.011308 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.011535 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl: configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.011600 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift podName:c7e65981-a172-41eb-ac4d-8bbe43248f0a nodeName:}" failed. No retries permitted until 2025-11-25 23:12:53.011578471 +0000 UTC m=+968.744924386 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift") pod "swift-proxy-6bd58cfcf7-dxxvl" (UID: "c7e65981-a172-41eb-ac4d-8bbe43248f0a") : configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.127711 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.127771 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.127810 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.128243 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.128296 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8" gracePeriod=600 Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.419021 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.419219 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.419233 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: E1125 23:12:51.419272 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift podName:ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad nodeName:}" failed. No retries permitted until 2025-11-25 23:12:55.419259259 +0000 UTC m=+971.152605094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift") pod "swift-storage-0" (UID: "ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad") : configmap "swift-ring-files" not found Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.546259 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8" exitCode=0 Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.546321 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8"} Nov 25 23:12:51 crc kubenswrapper[4761]: I1125 23:12:51.546352 4761 scope.go:117] "RemoveContainer" containerID="5e1af59ce3adc81d067b756b27b38ee5344cf5286ebf29d3fa1372d0281a31c4" Nov 25 23:12:53 crc kubenswrapper[4761]: I1125 23:12:53.046999 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:53 crc kubenswrapper[4761]: E1125 23:12:53.047216 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:53 crc kubenswrapper[4761]: E1125 23:12:53.047644 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl: configmap "swift-ring-files" not found Nov 25 23:12:53 crc kubenswrapper[4761]: E1125 23:12:53.047722 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift podName:c7e65981-a172-41eb-ac4d-8bbe43248f0a nodeName:}" failed. No retries permitted until 2025-11-25 23:12:57.047691071 +0000 UTC m=+972.781036896 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift") pod "swift-proxy-6bd58cfcf7-dxxvl" (UID: "c7e65981-a172-41eb-ac4d-8bbe43248f0a") : configmap "swift-ring-files" not found Nov 25 23:12:53 crc kubenswrapper[4761]: I1125 23:12:53.772496 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.378314 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-g5dpb"] Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.379234 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.388437 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-g5dpb"] Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.466967 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqtp7\" (UniqueName: \"kubernetes.io/projected/9b3bfb75-fcbe-4b6e-bb0e-8a09630901df-kube-api-access-hqtp7\") pod \"glance-operator-index-g5dpb\" (UID: \"9b3bfb75-fcbe-4b6e-bb0e-8a09630901df\") " pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.568575 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqtp7\" (UniqueName: \"kubernetes.io/projected/9b3bfb75-fcbe-4b6e-bb0e-8a09630901df-kube-api-access-hqtp7\") pod \"glance-operator-index-g5dpb\" (UID: \"9b3bfb75-fcbe-4b6e-bb0e-8a09630901df\") " pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.584208 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" event={"ID":"b067022b-e70b-4750-9d89-f9e48ca02b04","Type":"ContainerStarted","Data":"ede559326deb86939bbf907d362229433d79beee4d48b5ba19d83965447d904f"} Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.588795 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09"} Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.593803 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bbmjg" event={"ID":"327a4f2c-0d8d-4938-aa58-008bfa75bcca","Type":"ContainerStarted","Data":"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7"} Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.599344 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqtp7\" (UniqueName: \"kubernetes.io/projected/9b3bfb75-fcbe-4b6e-bb0e-8a09630901df-kube-api-access-hqtp7\") pod \"glance-operator-index-g5dpb\" (UID: \"9b3bfb75-fcbe-4b6e-bb0e-8a09630901df\") " pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.607959 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" podStartSLOduration=2.392271069 podStartE2EDuration="6.607936267s" podCreationTimestamp="2025-11-25 23:12:48 +0000 UTC" firstStartedPulling="2025-11-25 23:12:49.129735464 +0000 UTC m=+964.863081299" lastFinishedPulling="2025-11-25 23:12:53.345400662 +0000 UTC m=+969.078746497" observedRunningTime="2025-11-25 23:12:54.603062439 +0000 UTC m=+970.336408324" watchObservedRunningTime="2025-11-25 23:12:54.607936267 +0000 UTC m=+970.341282112" Nov 25 23:12:54 crc kubenswrapper[4761]: I1125 23:12:54.715237 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.313253 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-bbmjg" podStartSLOduration=2.689370234 podStartE2EDuration="6.313231664s" podCreationTimestamp="2025-11-25 23:12:49 +0000 UTC" firstStartedPulling="2025-11-25 23:12:49.783777932 +0000 UTC m=+965.517123767" lastFinishedPulling="2025-11-25 23:12:53.407639342 +0000 UTC m=+969.140985197" observedRunningTime="2025-11-25 23:12:54.644242683 +0000 UTC m=+970.377588548" watchObservedRunningTime="2025-11-25 23:12:55.313231664 +0000 UTC m=+971.046577519" Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.317044 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-g5dpb"] Nov 25 23:12:55 crc kubenswrapper[4761]: W1125 23:12:55.324825 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3bfb75_fcbe_4b6e_bb0e_8a09630901df.slice/crio-566c2dd46926479968dfbd749fa5742e1fb9d69cf0b45c244c38879ac432e539 WatchSource:0}: Error finding container 566c2dd46926479968dfbd749fa5742e1fb9d69cf0b45c244c38879ac432e539: Status 404 returned error can't find the container with id 566c2dd46926479968dfbd749fa5742e1fb9d69cf0b45c244c38879ac432e539 Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.480569 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:12:55 crc kubenswrapper[4761]: E1125 23:12:55.480861 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:55 crc kubenswrapper[4761]: E1125 23:12:55.480879 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 23:12:55 crc kubenswrapper[4761]: E1125 23:12:55.480925 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift podName:ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad nodeName:}" failed. No retries permitted until 2025-11-25 23:13:03.480909351 +0000 UTC m=+979.214255196 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift") pod "swift-storage-0" (UID: "ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad") : configmap "swift-ring-files" not found Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.601224 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-g5dpb" event={"ID":"9b3bfb75-fcbe-4b6e-bb0e-8a09630901df","Type":"ContainerStarted","Data":"847572d69de60c0f448fa3d4741efba9d7868bc986b64088a5a4f1dadc6a2c62"} Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.601314 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-g5dpb" event={"ID":"9b3bfb75-fcbe-4b6e-bb0e-8a09630901df","Type":"ContainerStarted","Data":"566c2dd46926479968dfbd749fa5742e1fb9d69cf0b45c244c38879ac432e539"} Nov 25 23:12:55 crc kubenswrapper[4761]: I1125 23:12:55.601409 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-bbmjg" podUID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" containerName="registry-server" containerID="cri-o://4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7" gracePeriod=2 Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.076809 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.108822 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-g5dpb" podStartSLOduration=2.011539258 podStartE2EDuration="2.108804419s" podCreationTimestamp="2025-11-25 23:12:54 +0000 UTC" firstStartedPulling="2025-11-25 23:12:55.332455891 +0000 UTC m=+971.065801726" lastFinishedPulling="2025-11-25 23:12:55.429721052 +0000 UTC m=+971.163066887" observedRunningTime="2025-11-25 23:12:55.621757791 +0000 UTC m=+971.355103626" watchObservedRunningTime="2025-11-25 23:12:56.108804419 +0000 UTC m=+971.842150254" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.190609 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8tsz\" (UniqueName: \"kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz\") pod \"327a4f2c-0d8d-4938-aa58-008bfa75bcca\" (UID: \"327a4f2c-0d8d-4938-aa58-008bfa75bcca\") " Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.210956 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz" (OuterVolumeSpecName: "kube-api-access-z8tsz") pod "327a4f2c-0d8d-4938-aa58-008bfa75bcca" (UID: "327a4f2c-0d8d-4938-aa58-008bfa75bcca"). InnerVolumeSpecName "kube-api-access-z8tsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.293260 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8tsz\" (UniqueName: \"kubernetes.io/projected/327a4f2c-0d8d-4938-aa58-008bfa75bcca-kube-api-access-z8tsz\") on node \"crc\" DevicePath \"\"" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.612967 4761 generic.go:334] "Generic (PLEG): container finished" podID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" containerID="4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7" exitCode=0 Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.613038 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bbmjg" event={"ID":"327a4f2c-0d8d-4938-aa58-008bfa75bcca","Type":"ContainerDied","Data":"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7"} Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.613098 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bbmjg" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.613120 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bbmjg" event={"ID":"327a4f2c-0d8d-4938-aa58-008bfa75bcca","Type":"ContainerDied","Data":"311988575ce39513eec3008af25a524c1ff36a30913b661dc51f7fa9072122c8"} Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.613148 4761 scope.go:117] "RemoveContainer" containerID="4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.661009 4761 scope.go:117] "RemoveContainer" containerID="4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7" Nov 25 23:12:56 crc kubenswrapper[4761]: E1125 23:12:56.661924 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7\": container with ID starting with 4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7 not found: ID does not exist" containerID="4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.661954 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7"} err="failed to get container status \"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7\": rpc error: code = NotFound desc = could not find container \"4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7\": container with ID starting with 4563f2737d0fddac273fedd29a43bcc13531b64c0f4282cb6a8c98e5f4ff91e7 not found: ID does not exist" Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.673541 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:56 crc kubenswrapper[4761]: I1125 23:12:56.680276 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-bbmjg"] Nov 25 23:12:57 crc kubenswrapper[4761]: I1125 23:12:57.025563 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" path="/var/lib/kubelet/pods/327a4f2c-0d8d-4938-aa58-008bfa75bcca/volumes" Nov 25 23:12:57 crc kubenswrapper[4761]: I1125 23:12:57.106447 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:12:57 crc kubenswrapper[4761]: E1125 23:12:57.106626 4761 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 23:12:57 crc kubenswrapper[4761]: E1125 23:12:57.106659 4761 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl: configmap "swift-ring-files" not found Nov 25 23:12:57 crc kubenswrapper[4761]: E1125 23:12:57.106762 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift podName:c7e65981-a172-41eb-ac4d-8bbe43248f0a nodeName:}" failed. No retries permitted until 2025-11-25 23:13:05.106735264 +0000 UTC m=+980.840081149 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift") pod "swift-proxy-6bd58cfcf7-dxxvl" (UID: "c7e65981-a172-41eb-ac4d-8bbe43248f0a") : configmap "swift-ring-files" not found Nov 25 23:13:00 crc kubenswrapper[4761]: I1125 23:13:00.644287 4761 generic.go:334] "Generic (PLEG): container finished" podID="b067022b-e70b-4750-9d89-f9e48ca02b04" containerID="ede559326deb86939bbf907d362229433d79beee4d48b5ba19d83965447d904f" exitCode=0 Nov 25 23:13:00 crc kubenswrapper[4761]: I1125 23:13:00.644340 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" event={"ID":"b067022b-e70b-4750-9d89-f9e48ca02b04","Type":"ContainerDied","Data":"ede559326deb86939bbf907d362229433d79beee4d48b5ba19d83965447d904f"} Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.032988 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087066 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fz6w\" (UniqueName: \"kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087215 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087370 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087407 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087505 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.087600 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf\") pod \"b067022b-e70b-4750-9d89-f9e48ca02b04\" (UID: \"b067022b-e70b-4750-9d89-f9e48ca02b04\") " Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.088637 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.089810 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.092619 4761 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.092667 4761 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b067022b-e70b-4750-9d89-f9e48ca02b04-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.095018 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w" (OuterVolumeSpecName: "kube-api-access-8fz6w") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "kube-api-access-8fz6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.101775 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.114966 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.122284 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts" (OuterVolumeSpecName: "scripts") pod "b067022b-e70b-4750-9d89-f9e48ca02b04" (UID: "b067022b-e70b-4750-9d89-f9e48ca02b04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.194483 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fz6w\" (UniqueName: \"kubernetes.io/projected/b067022b-e70b-4750-9d89-f9e48ca02b04-kube-api-access-8fz6w\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.194538 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b067022b-e70b-4750-9d89-f9e48ca02b04-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.194557 4761 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.194575 4761 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b067022b-e70b-4750-9d89-f9e48ca02b04-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.664310 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" event={"ID":"b067022b-e70b-4750-9d89-f9e48ca02b04","Type":"ContainerDied","Data":"5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf"} Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.664361 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5336ff9ae4ce49ad887ddbf4626c5628bce6cdbc9736fb537f497588c15ca4cf" Nov 25 23:13:02 crc kubenswrapper[4761]: I1125 23:13:02.664418 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-9mc86" Nov 25 23:13:03 crc kubenswrapper[4761]: I1125 23:13:03.515667 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:13:03 crc kubenswrapper[4761]: I1125 23:13:03.536909 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad-etc-swift\") pod \"swift-storage-0\" (UID: \"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:13:03 crc kubenswrapper[4761]: I1125 23:13:03.550916 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 23:13:04 crc kubenswrapper[4761]: I1125 23:13:04.079995 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 23:13:04 crc kubenswrapper[4761]: I1125 23:13:04.705871 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"1ef2ccba47ab86c840e8380af871d84ddaf764cdcbbcf790bdf50075ed6f2feb"} Nov 25 23:13:04 crc kubenswrapper[4761]: I1125 23:13:04.715456 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:13:04 crc kubenswrapper[4761]: I1125 23:13:04.715546 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:13:04 crc kubenswrapper[4761]: I1125 23:13:04.757386 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.140708 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.147191 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7e65981-a172-41eb-ac4d-8bbe43248f0a-etc-swift\") pod \"swift-proxy-6bd58cfcf7-dxxvl\" (UID: \"c7e65981-a172-41eb-ac4d-8bbe43248f0a\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.206147 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.669416 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl"] Nov 25 23:13:05 crc kubenswrapper[4761]: W1125 23:13:05.675375 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7e65981_a172_41eb_ac4d_8bbe43248f0a.slice/crio-6bd3eca7aad206e057072eefd4ea1a8fbbdc951ae886063e6a0ff3085d72d2fc WatchSource:0}: Error finding container 6bd3eca7aad206e057072eefd4ea1a8fbbdc951ae886063e6a0ff3085d72d2fc: Status 404 returned error can't find the container with id 6bd3eca7aad206e057072eefd4ea1a8fbbdc951ae886063e6a0ff3085d72d2fc Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.716241 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"6226acd0a1fc4299082689db8b19d005de83707971ad75665aec23350f14527a"} Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.716317 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"921c0c76d4e6aad00fa6cab1563f83a6a2867ccf722d0fddb7f265712e34d690"} Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.720233 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" event={"ID":"c7e65981-a172-41eb-ac4d-8bbe43248f0a","Type":"ContainerStarted","Data":"6bd3eca7aad206e057072eefd4ea1a8fbbdc951ae886063e6a0ff3085d72d2fc"} Nov 25 23:13:05 crc kubenswrapper[4761]: I1125 23:13:05.764066 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-g5dpb" Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.734928 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"6cc951dd53c814a61ab6384cea5c6908174db746dfca13775615667e4039ccd8"} Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.735264 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"a4d5f39901eb506338d7c912d97142132f90553d5277a04207ab361695db022c"} Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.738468 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" event={"ID":"c7e65981-a172-41eb-ac4d-8bbe43248f0a","Type":"ContainerStarted","Data":"268ca1f3020b8cb064f89fc93edb9834d9dd0decef92ba78247f7eb611114ee0"} Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.738538 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" event={"ID":"c7e65981-a172-41eb-ac4d-8bbe43248f0a","Type":"ContainerStarted","Data":"2f0241376ab57e7f2caf5d57ecc43d2447a3b7529fb4c525ae684c5cdba672bb"} Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.738650 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:06 crc kubenswrapper[4761]: I1125 23:13:06.762320 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" podStartSLOduration=17.762302796 podStartE2EDuration="17.762302796s" podCreationTimestamp="2025-11-25 23:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:13:06.761234378 +0000 UTC m=+982.494580253" watchObservedRunningTime="2025-11-25 23:13:06.762302796 +0000 UTC m=+982.495648631" Nov 25 23:13:07 crc kubenswrapper[4761]: I1125 23:13:07.757855 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"70ed28162c8f59d6dd70f223564eba32820629b665d843780c06f026b2fbe8a1"} Nov 25 23:13:07 crc kubenswrapper[4761]: I1125 23:13:07.758173 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:07 crc kubenswrapper[4761]: I1125 23:13:07.758191 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"ef12085ad4b577e7be2c33093fd32f3f2fa52800a8571d09a24ded55fe6cf325"} Nov 25 23:13:07 crc kubenswrapper[4761]: I1125 23:13:07.758204 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"47fb61e6f883f0c10b39654471da387786f55ff3a3d2db045e400fa67f9552fe"} Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.433440 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq"] Nov 25 23:13:08 crc kubenswrapper[4761]: E1125 23:13:08.434159 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b067022b-e70b-4750-9d89-f9e48ca02b04" containerName="swift-ring-rebalance" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.434182 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b067022b-e70b-4750-9d89-f9e48ca02b04" containerName="swift-ring-rebalance" Nov 25 23:13:08 crc kubenswrapper[4761]: E1125 23:13:08.434214 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" containerName="registry-server" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.434225 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" containerName="registry-server" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.434432 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b067022b-e70b-4750-9d89-f9e48ca02b04" containerName="swift-ring-rebalance" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.434457 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="327a4f2c-0d8d-4938-aa58-008bfa75bcca" containerName="registry-server" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.435857 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.438974 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pp97f" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.441479 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq"] Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.495638 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7txt\" (UniqueName: \"kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.495719 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.495754 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.596955 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7txt\" (UniqueName: \"kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.597010 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.597033 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.597445 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.597911 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.620942 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7txt\" (UniqueName: \"kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt\") pod \"06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.760535 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:08 crc kubenswrapper[4761]: I1125 23:13:08.770964 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"30ecb6840e60a74d6886f1f69efa760832d2901e540aafae74fd72be13ba55b3"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.569220 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq"] Nov 25 23:13:09 crc kubenswrapper[4761]: W1125 23:13:09.578782 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52cc79c4_6ed7_463c_afae_9e4e9740ddbb.slice/crio-2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54 WatchSource:0}: Error finding container 2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54: Status 404 returned error can't find the container with id 2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54 Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.786004 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"cda904732e49711b5ba2bf9c24312135ce67dd1a8bfb04dc424ce04824ea3052"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.786382 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"f5784768d00394a07f56419c4ab671a64d64f2759a255dea7350f6d4f156e5d9"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.786395 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"828c8abdb839169c8960e30781e9d6e679a8b09a9fd6e5a78bc00e1989bbb7cf"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.786406 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"bdf02ee16ee6e199b82e5c9729c27ab026c1871c1601ee0c417ba9086a6f7009"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.787788 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerStarted","Data":"9dc5be0fa303758a0dad8abfe272912c8f119ab85efa05292db2d0efc01ef34b"} Nov 25 23:13:09 crc kubenswrapper[4761]: I1125 23:13:09.787821 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerStarted","Data":"2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54"} Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.215647 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.216468 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-dxxvl" Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.805827 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"9f95e0ec7b3b5be2af21c4000f56a5a9d6c33abd9f9be95faa6d30f4648a7bab"} Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.806183 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"228c88095b04d038c2eb623bbb230da04fddf56530cb6201f75a413928b46a48"} Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.806197 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad","Type":"ContainerStarted","Data":"af012f4c1c90dcde1f7ef6800e8b6c3d606be3a6d6cdc70417e9613920959a47"} Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.808252 4761 generic.go:334] "Generic (PLEG): container finished" podID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerID="9dc5be0fa303758a0dad8abfe272912c8f119ab85efa05292db2d0efc01ef34b" exitCode=0 Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.808437 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerDied","Data":"9dc5be0fa303758a0dad8abfe272912c8f119ab85efa05292db2d0efc01ef34b"} Nov 25 23:13:10 crc kubenswrapper[4761]: I1125 23:13:10.861515 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.823267874 podStartE2EDuration="24.861491567s" podCreationTimestamp="2025-11-25 23:12:46 +0000 UTC" firstStartedPulling="2025-11-25 23:13:04.100759603 +0000 UTC m=+979.834105468" lastFinishedPulling="2025-11-25 23:13:09.138983326 +0000 UTC m=+984.872329161" observedRunningTime="2025-11-25 23:13:10.849499721 +0000 UTC m=+986.582845606" watchObservedRunningTime="2025-11-25 23:13:10.861491567 +0000 UTC m=+986.594837412" Nov 25 23:13:11 crc kubenswrapper[4761]: I1125 23:13:11.820995 4761 generic.go:334] "Generic (PLEG): container finished" podID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerID="b6879dbf2e3152d36947cf0c8a09bac109b4fee1e86845bdfdbc2af7330c7967" exitCode=0 Nov 25 23:13:11 crc kubenswrapper[4761]: I1125 23:13:11.821114 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerDied","Data":"b6879dbf2e3152d36947cf0c8a09bac109b4fee1e86845bdfdbc2af7330c7967"} Nov 25 23:13:12 crc kubenswrapper[4761]: I1125 23:13:12.834076 4761 generic.go:334] "Generic (PLEG): container finished" podID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerID="0b934f4edbff3629c033adc018eb440fd14b23d709f1fa5999c829533ad7234f" exitCode=0 Nov 25 23:13:12 crc kubenswrapper[4761]: I1125 23:13:12.834125 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerDied","Data":"0b934f4edbff3629c033adc018eb440fd14b23d709f1fa5999c829533ad7234f"} Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.152798 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.185327 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle\") pod \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.185449 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util\") pod \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.185590 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7txt\" (UniqueName: \"kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt\") pod \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\" (UID: \"52cc79c4-6ed7-463c-afae-9e4e9740ddbb\") " Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.187496 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle" (OuterVolumeSpecName: "bundle") pod "52cc79c4-6ed7-463c-afae-9e4e9740ddbb" (UID: "52cc79c4-6ed7-463c-afae-9e4e9740ddbb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.197069 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt" (OuterVolumeSpecName: "kube-api-access-q7txt") pod "52cc79c4-6ed7-463c-afae-9e4e9740ddbb" (UID: "52cc79c4-6ed7-463c-afae-9e4e9740ddbb"). InnerVolumeSpecName "kube-api-access-q7txt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.203123 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util" (OuterVolumeSpecName: "util") pod "52cc79c4-6ed7-463c-afae-9e4e9740ddbb" (UID: "52cc79c4-6ed7-463c-afae-9e4e9740ddbb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.287794 4761 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.287860 4761 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-util\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.287882 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7txt\" (UniqueName: \"kubernetes.io/projected/52cc79c4-6ed7-463c-afae-9e4e9740ddbb-kube-api-access-q7txt\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.856186 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" event={"ID":"52cc79c4-6ed7-463c-afae-9e4e9740ddbb","Type":"ContainerDied","Data":"2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54"} Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.856233 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fa3de4c62a0013e88a80a7de9f9b9affea7f8555f76e8daafd735e2f4132e54" Nov 25 23:13:14 crc kubenswrapper[4761]: I1125 23:13:14.856318 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.279733 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp"] Nov 25 23:13:29 crc kubenswrapper[4761]: E1125 23:13:29.280639 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="util" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.280654 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="util" Nov 25 23:13:29 crc kubenswrapper[4761]: E1125 23:13:29.280673 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="extract" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.280680 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="extract" Nov 25 23:13:29 crc kubenswrapper[4761]: E1125 23:13:29.280722 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="pull" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.280731 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="pull" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.280875 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="52cc79c4-6ed7-463c-afae-9e4e9740ddbb" containerName="extract" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.281360 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.283215 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pvqmg" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.290499 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.292825 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp"] Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.409732 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-webhook-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.409798 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8vl\" (UniqueName: \"kubernetes.io/projected/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-kube-api-access-rt8vl\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.409913 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-apiservice-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.510948 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8vl\" (UniqueName: \"kubernetes.io/projected/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-kube-api-access-rt8vl\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.511226 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-apiservice-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.511370 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-webhook-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.518903 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-webhook-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.519086 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-apiservice-cert\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.530645 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8vl\" (UniqueName: \"kubernetes.io/projected/3daf70a9-4dab-4ace-82c0-9c7e48512cf6-kube-api-access-rt8vl\") pod \"glance-operator-controller-manager-66b769fbc8-xf6tp\" (UID: \"3daf70a9-4dab-4ace-82c0-9c7e48512cf6\") " pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:29 crc kubenswrapper[4761]: I1125 23:13:29.603827 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:30 crc kubenswrapper[4761]: I1125 23:13:30.115918 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp"] Nov 25 23:13:30 crc kubenswrapper[4761]: I1125 23:13:30.980003 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerStarted","Data":"820535c6fb3a86c40b17774791eefac2a9dca7999fd31a9d87292518dd2a7d18"} Nov 25 23:13:31 crc kubenswrapper[4761]: I1125 23:13:31.990487 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerStarted","Data":"2ba5de066a343f8a191d00b472cce29283efeea1f98421fb4d1e5a4a3e7e0fb9"} Nov 25 23:13:31 crc kubenswrapper[4761]: I1125 23:13:31.990806 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:32 crc kubenswrapper[4761]: I1125 23:13:32.013653 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" podStartSLOduration=1.6378204950000002 podStartE2EDuration="3.013628923s" podCreationTimestamp="2025-11-25 23:13:29 +0000 UTC" firstStartedPulling="2025-11-25 23:13:30.11446848 +0000 UTC m=+1005.847814315" lastFinishedPulling="2025-11-25 23:13:31.490276908 +0000 UTC m=+1007.223622743" observedRunningTime="2025-11-25 23:13:32.010780158 +0000 UTC m=+1007.744126023" watchObservedRunningTime="2025-11-25 23:13:32.013628923 +0000 UTC m=+1007.746974768" Nov 25 23:13:39 crc kubenswrapper[4761]: I1125 23:13:39.608363 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.746213 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.748243 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.752924 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.753345 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-lmqzr" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.753524 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.753631 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.755768 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-kv6cr"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.756826 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.769679 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-89b9-account-create-update-nk6dv"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.771055 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.781273 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.783503 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.792354 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-89b9-account-create-update-nk6dv"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.794286 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-kv6cr"] Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912464 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n99d8\" (UniqueName: \"kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912540 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2cb5\" (UniqueName: \"kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912589 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912622 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912649 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912695 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vpw\" (UniqueName: \"kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912754 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:42 crc kubenswrapper[4761]: I1125 23:13:42.912776 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013781 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013830 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013872 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vpw\" (UniqueName: \"kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013905 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013923 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013963 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n99d8\" (UniqueName: \"kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.013993 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2cb5\" (UniqueName: \"kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.014807 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.014766 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.014761 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.014936 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.015361 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.037392 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.041332 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vpw\" (UniqueName: \"kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw\") pod \"glance-89b9-account-create-update-nk6dv\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.041387 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n99d8\" (UniqueName: \"kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8\") pod \"glance-db-create-kv6cr\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.043897 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2cb5\" (UniqueName: \"kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5\") pod \"openstackclient\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.070933 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.121489 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.133050 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.535819 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:13:43 crc kubenswrapper[4761]: W1125 23:13:43.536169 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13642318_226b_4cf1_84fc_38d0eb5bd797.slice/crio-85c8d5d2c484928f99bf365502f2360d8a3ce589a37e37b971f537122c13fb6a WatchSource:0}: Error finding container 85c8d5d2c484928f99bf365502f2360d8a3ce589a37e37b971f537122c13fb6a: Status 404 returned error can't find the container with id 85c8d5d2c484928f99bf365502f2360d8a3ce589a37e37b971f537122c13fb6a Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.612820 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-kv6cr"] Nov 25 23:13:43 crc kubenswrapper[4761]: W1125 23:13:43.619321 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a06b3e4_da6e_40a5_bec8_3fcd8aefceae.slice/crio-d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67 WatchSource:0}: Error finding container d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67: Status 404 returned error can't find the container with id d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67 Nov 25 23:13:43 crc kubenswrapper[4761]: I1125 23:13:43.661261 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-89b9-account-create-update-nk6dv"] Nov 25 23:13:43 crc kubenswrapper[4761]: W1125 23:13:43.669908 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod791ea0cf_ae15_4e32_911d_93c65978db8d.slice/crio-da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41 WatchSource:0}: Error finding container da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41: Status 404 returned error can't find the container with id da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41 Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.104140 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"13642318-226b-4cf1-84fc-38d0eb5bd797","Type":"ContainerStarted","Data":"85c8d5d2c484928f99bf365502f2360d8a3ce589a37e37b971f537122c13fb6a"} Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.106304 4761 generic.go:334] "Generic (PLEG): container finished" podID="791ea0cf-ae15-4e32-911d-93c65978db8d" containerID="9633597078bb507c73766fef261263383998d4db894a50ee8c63b8e499c59573" exitCode=0 Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.106384 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" event={"ID":"791ea0cf-ae15-4e32-911d-93c65978db8d","Type":"ContainerDied","Data":"9633597078bb507c73766fef261263383998d4db894a50ee8c63b8e499c59573"} Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.106411 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" event={"ID":"791ea0cf-ae15-4e32-911d-93c65978db8d","Type":"ContainerStarted","Data":"da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41"} Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.109229 4761 generic.go:334] "Generic (PLEG): container finished" podID="3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" containerID="ce50da77323e94f7c89db224b8eb05c4265e5f6d6a1bb23b2c689baad23a602b" exitCode=0 Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.109270 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kv6cr" event={"ID":"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae","Type":"ContainerDied","Data":"ce50da77323e94f7c89db224b8eb05c4265e5f6d6a1bb23b2c689baad23a602b"} Nov 25 23:13:44 crc kubenswrapper[4761]: I1125 23:13:44.109294 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kv6cr" event={"ID":"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae","Type":"ContainerStarted","Data":"d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67"} Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.479898 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.485436 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.556660 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts\") pod \"791ea0cf-ae15-4e32-911d-93c65978db8d\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.556769 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts\") pod \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.556791 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7vpw\" (UniqueName: \"kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw\") pod \"791ea0cf-ae15-4e32-911d-93c65978db8d\" (UID: \"791ea0cf-ae15-4e32-911d-93c65978db8d\") " Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.556847 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n99d8\" (UniqueName: \"kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8\") pod \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\" (UID: \"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae\") " Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.557305 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "791ea0cf-ae15-4e32-911d-93c65978db8d" (UID: "791ea0cf-ae15-4e32-911d-93c65978db8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.557557 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" (UID: "3a06b3e4-da6e-40a5-bec8-3fcd8aefceae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.564335 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw" (OuterVolumeSpecName: "kube-api-access-f7vpw") pod "791ea0cf-ae15-4e32-911d-93c65978db8d" (UID: "791ea0cf-ae15-4e32-911d-93c65978db8d"). InnerVolumeSpecName "kube-api-access-f7vpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.564906 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8" (OuterVolumeSpecName: "kube-api-access-n99d8") pod "3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" (UID: "3a06b3e4-da6e-40a5-bec8-3fcd8aefceae"). InnerVolumeSpecName "kube-api-access-n99d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.658990 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.659044 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7vpw\" (UniqueName: \"kubernetes.io/projected/791ea0cf-ae15-4e32-911d-93c65978db8d-kube-api-access-f7vpw\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.659056 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n99d8\" (UniqueName: \"kubernetes.io/projected/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae-kube-api-access-n99d8\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:45 crc kubenswrapper[4761]: I1125 23:13:45.659065 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/791ea0cf-ae15-4e32-911d-93c65978db8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.130809 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" event={"ID":"791ea0cf-ae15-4e32-911d-93c65978db8d","Type":"ContainerDied","Data":"da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41"} Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.130862 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da291ddc3abf692bb30a9d3474aedeb99ec5b4f248345fe0544fe6894a235e41" Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.130887 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-89b9-account-create-update-nk6dv" Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.148941 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kv6cr" event={"ID":"3a06b3e4-da6e-40a5-bec8-3fcd8aefceae","Type":"ContainerDied","Data":"d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67"} Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.149020 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0d13bb429936c81373ba7b20ed78f8879406b25cd8c33d4efede54782b97c67" Nov 25 23:13:46 crc kubenswrapper[4761]: I1125 23:13:46.149259 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kv6cr" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.041664 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-jw7qd"] Nov 25 23:13:48 crc kubenswrapper[4761]: E1125 23:13:48.043026 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791ea0cf-ae15-4e32-911d-93c65978db8d" containerName="mariadb-account-create-update" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.043093 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="791ea0cf-ae15-4e32-911d-93c65978db8d" containerName="mariadb-account-create-update" Nov 25 23:13:48 crc kubenswrapper[4761]: E1125 23:13:48.043157 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" containerName="mariadb-database-create" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.043211 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" containerName="mariadb-database-create" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.043394 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="791ea0cf-ae15-4e32-911d-93c65978db8d" containerName="mariadb-account-create-update" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.043453 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" containerName="mariadb-database-create" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.043911 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.046891 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-ncqfx" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.046933 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.055171 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-jw7qd"] Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.203354 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2g4v\" (UniqueName: \"kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.203403 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.203426 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.305222 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2g4v\" (UniqueName: \"kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.305273 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.305298 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.313445 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.313523 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.327465 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2g4v\" (UniqueName: \"kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v\") pod \"glance-db-sync-jw7qd\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:48 crc kubenswrapper[4761]: I1125 23:13:48.360646 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:13:52 crc kubenswrapper[4761]: I1125 23:13:52.002045 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-jw7qd"] Nov 25 23:13:52 crc kubenswrapper[4761]: W1125 23:13:52.006779 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d87d1f9_7c4a_4dab_bee8_6f20f4a02415.slice/crio-debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a WatchSource:0}: Error finding container debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a: Status 404 returned error can't find the container with id debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a Nov 25 23:13:52 crc kubenswrapper[4761]: I1125 23:13:52.216840 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-jw7qd" event={"ID":"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415","Type":"ContainerStarted","Data":"debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a"} Nov 25 23:13:52 crc kubenswrapper[4761]: I1125 23:13:52.219414 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"13642318-226b-4cf1-84fc-38d0eb5bd797","Type":"ContainerStarted","Data":"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d"} Nov 25 23:13:52 crc kubenswrapper[4761]: I1125 23:13:52.236968 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.165148988 podStartE2EDuration="10.236947646s" podCreationTimestamp="2025-11-25 23:13:42 +0000 UTC" firstStartedPulling="2025-11-25 23:13:43.539225322 +0000 UTC m=+1019.272571167" lastFinishedPulling="2025-11-25 23:13:51.61102399 +0000 UTC m=+1027.344369825" observedRunningTime="2025-11-25 23:13:52.236380271 +0000 UTC m=+1027.969726136" watchObservedRunningTime="2025-11-25 23:13:52.236947646 +0000 UTC m=+1027.970293481" Nov 25 23:14:07 crc kubenswrapper[4761]: I1125 23:14:07.374841 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-jw7qd" event={"ID":"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415","Type":"ContainerStarted","Data":"bd97a4391c11a978c36a4883f1ce7ade43cf0bcd91e22909f420dfb79e198787"} Nov 25 23:14:07 crc kubenswrapper[4761]: I1125 23:14:07.403154 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-jw7qd" podStartSLOduration=6.8275688 podStartE2EDuration="19.403128435s" podCreationTimestamp="2025-11-25 23:13:48 +0000 UTC" firstStartedPulling="2025-11-25 23:13:52.009126325 +0000 UTC m=+1027.742472160" lastFinishedPulling="2025-11-25 23:14:04.58468595 +0000 UTC m=+1040.318031795" observedRunningTime="2025-11-25 23:14:07.39501447 +0000 UTC m=+1043.128360345" watchObservedRunningTime="2025-11-25 23:14:07.403128435 +0000 UTC m=+1043.136474310" Nov 25 23:14:14 crc kubenswrapper[4761]: I1125 23:14:14.436768 4761 generic.go:334] "Generic (PLEG): container finished" podID="1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" containerID="bd97a4391c11a978c36a4883f1ce7ade43cf0bcd91e22909f420dfb79e198787" exitCode=0 Nov 25 23:14:14 crc kubenswrapper[4761]: I1125 23:14:14.436867 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-jw7qd" event={"ID":"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415","Type":"ContainerDied","Data":"bd97a4391c11a978c36a4883f1ce7ade43cf0bcd91e22909f420dfb79e198787"} Nov 25 23:14:15 crc kubenswrapper[4761]: I1125 23:14:15.865787 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.040153 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2g4v\" (UniqueName: \"kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v\") pod \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.040219 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data\") pod \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.040301 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data\") pod \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\" (UID: \"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415\") " Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.046086 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" (UID: "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.049478 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v" (OuterVolumeSpecName: "kube-api-access-z2g4v") pod "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" (UID: "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415"). InnerVolumeSpecName "kube-api-access-z2g4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.101022 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data" (OuterVolumeSpecName: "config-data") pod "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" (UID: "1d87d1f9-7c4a-4dab-bee8-6f20f4a02415"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.142069 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2g4v\" (UniqueName: \"kubernetes.io/projected/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-kube-api-access-z2g4v\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.142098 4761 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.142131 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.459205 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-jw7qd" event={"ID":"1d87d1f9-7c4a-4dab-bee8-6f20f4a02415","Type":"ContainerDied","Data":"debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a"} Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.459263 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="debdfd5b0a7bc657d83f4d495d658b344596eec0848689cf4a1e3a48e5f5182a" Nov 25 23:14:16 crc kubenswrapper[4761]: I1125 23:14:16.459318 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-jw7qd" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.857691 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:14:17 crc kubenswrapper[4761]: E1125 23:14:17.858986 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" containerName="glance-db-sync" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.859084 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" containerName="glance-db-sync" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.859351 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" containerName="glance-db-sync" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.860322 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.862834 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.863189 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.863559 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-ncqfx" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.874185 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.956182 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.957773 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966744 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966805 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966843 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g87th\" (UniqueName: \"kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966874 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966905 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966924 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966953 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966979 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.966996 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.967019 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.967039 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.967075 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.967102 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.967126 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:17 crc kubenswrapper[4761]: I1125 23:14:17.969964 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.068861 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.068929 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.068952 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.068984 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.068998 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069048 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069128 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069171 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069192 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069302 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069367 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g87th\" (UniqueName: \"kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069428 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069488 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069559 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069585 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069633 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069650 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069651 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069649 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069681 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069767 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069829 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069850 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069867 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069876 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069881 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069928 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069932 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069957 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069978 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.069994 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070036 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt9qg\" (UniqueName: \"kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070075 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070132 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070158 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070234 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070257 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070243 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.070550 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.088123 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.091444 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g87th\" (UniqueName: \"kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.093208 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.095600 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.096398 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172001 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172046 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172071 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172089 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172103 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172128 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt9qg\" (UniqueName: \"kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172164 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172159 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172192 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172231 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172265 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172286 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172306 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172323 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172365 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172274 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172403 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172409 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172450 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172360 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172269 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172684 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.172495 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.173193 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.173302 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.179093 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.180417 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.204172 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.207794 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt9qg\" (UniqueName: \"kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.212936 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.233182 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.274938 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.518011 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:18 crc kubenswrapper[4761]: I1125 23:14:18.687811 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:14:18 crc kubenswrapper[4761]: W1125 23:14:18.691805 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod660e9154_b8d1_468e_bbe5_bc884c4ad0a7.slice/crio-0379144f637967767bf0c61c3dc7469b8019fa7c1b2df88f5d993036a0e5cf16 WatchSource:0}: Error finding container 0379144f637967767bf0c61c3dc7469b8019fa7c1b2df88f5d993036a0e5cf16: Status 404 returned error can't find the container with id 0379144f637967767bf0c61c3dc7469b8019fa7c1b2df88f5d993036a0e5cf16 Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.487414 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerStarted","Data":"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.488109 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerStarted","Data":"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.488133 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerStarted","Data":"4b36d08d176cbcd881e48fe8ef4a3582630cffeca7dfd97d416ae462a7c101cf"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.493761 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerStarted","Data":"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.493813 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerStarted","Data":"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.493833 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerStarted","Data":"0379144f637967767bf0c61c3dc7469b8019fa7c1b2df88f5d993036a0e5cf16"} Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.543814 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.543778962 podStartE2EDuration="3.543778962s" podCreationTimestamp="2025-11-25 23:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:14:19.527166661 +0000 UTC m=+1055.260512596" watchObservedRunningTime="2025-11-25 23:14:19.543778962 +0000 UTC m=+1055.277124867" Nov 25 23:14:19 crc kubenswrapper[4761]: I1125 23:14:19.575598 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.575570726 podStartE2EDuration="2.575570726s" podCreationTimestamp="2025-11-25 23:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:14:19.563334341 +0000 UTC m=+1055.296680206" watchObservedRunningTime="2025-11-25 23:14:19.575570726 +0000 UTC m=+1055.308916591" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.233847 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.234615 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.275296 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.276146 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.276196 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.287357 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.325278 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.325521 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.577741 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.577787 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.577801 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:28 crc kubenswrapper[4761]: I1125 23:14:28.577813 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.424434 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.449511 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.676131 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.676552 4761 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.679861 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:14:30 crc kubenswrapper[4761]: I1125 23:14:30.734549 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:32 crc kubenswrapper[4761]: I1125 23:14:32.610060 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-httpd" containerID="cri-o://ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41" gracePeriod=30 Nov 25 23:14:32 crc kubenswrapper[4761]: I1125 23:14:32.610312 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-log" containerID="cri-o://9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846" gracePeriod=30 Nov 25 23:14:33 crc kubenswrapper[4761]: I1125 23:14:33.619087 4761 generic.go:334] "Generic (PLEG): container finished" podID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerID="9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846" exitCode=143 Nov 25 23:14:33 crc kubenswrapper[4761]: I1125 23:14:33.619179 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerDied","Data":"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846"} Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.224982 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404685 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404797 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404850 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404893 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt9qg\" (UniqueName: \"kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404936 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.404970 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405047 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405100 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405156 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405188 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405208 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405273 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405225 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405331 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev" (OuterVolumeSpecName: "dev") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405218 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405428 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405461 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405502 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405519 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run\") pod \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\" (UID: \"310ccbc5-38e7-402c-8cc0-adf9c7038e42\") " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.405955 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys" (OuterVolumeSpecName: "sys") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406006 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406048 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run" (OuterVolumeSpecName: "run") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406265 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406297 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406312 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406327 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406343 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406356 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406369 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406384 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/310ccbc5-38e7-402c-8cc0-adf9c7038e42-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.406961 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs" (OuterVolumeSpecName: "logs") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.411524 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg" (OuterVolumeSpecName: "kube-api-access-mt9qg") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "kube-api-access-mt9qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.412642 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.413097 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts" (OuterVolumeSpecName: "scripts") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.415482 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.461214 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data" (OuterVolumeSpecName: "config-data") pod "310ccbc5-38e7-402c-8cc0-adf9c7038e42" (UID: "310ccbc5-38e7-402c-8cc0-adf9c7038e42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507475 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507510 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310ccbc5-38e7-402c-8cc0-adf9c7038e42-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507523 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt9qg\" (UniqueName: \"kubernetes.io/projected/310ccbc5-38e7-402c-8cc0-adf9c7038e42-kube-api-access-mt9qg\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507542 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/310ccbc5-38e7-402c-8cc0-adf9c7038e42-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507587 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.507604 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.520580 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.524228 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.609349 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.609410 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.655442 4761 generic.go:334] "Generic (PLEG): container finished" podID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerID="ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41" exitCode=0 Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.655515 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.655511 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerDied","Data":"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41"} Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.655596 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"310ccbc5-38e7-402c-8cc0-adf9c7038e42","Type":"ContainerDied","Data":"4b36d08d176cbcd881e48fe8ef4a3582630cffeca7dfd97d416ae462a7c101cf"} Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.655636 4761 scope.go:117] "RemoveContainer" containerID="ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.675689 4761 scope.go:117] "RemoveContainer" containerID="9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.696475 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.707972 4761 scope.go:117] "RemoveContainer" containerID="ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.711220 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:36 crc kubenswrapper[4761]: E1125 23:14:36.713193 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41\": container with ID starting with ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41 not found: ID does not exist" containerID="ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.713234 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41"} err="failed to get container status \"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41\": rpc error: code = NotFound desc = could not find container \"ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41\": container with ID starting with ff6e8f7e7ae74e7fb2a7fc6809a5665b7b5cf2ddca2a0595c132f1a3b7f05a41 not found: ID does not exist" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.713260 4761 scope.go:117] "RemoveContainer" containerID="9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846" Nov 25 23:14:36 crc kubenswrapper[4761]: E1125 23:14:36.714608 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846\": container with ID starting with 9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846 not found: ID does not exist" containerID="9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.714632 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846"} err="failed to get container status \"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846\": rpc error: code = NotFound desc = could not find container \"9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846\": container with ID starting with 9adc76fabb90aecafe24a0a155368141631a4cb8db02802e925bf15d40ea2846 not found: ID does not exist" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.728197 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:36 crc kubenswrapper[4761]: E1125 23:14:36.728757 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-httpd" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.728784 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-httpd" Nov 25 23:14:36 crc kubenswrapper[4761]: E1125 23:14:36.728819 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-log" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.728831 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-log" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.729095 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-httpd" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.729131 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" containerName="glance-log" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.730737 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.744437 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914419 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914511 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914570 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914601 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914632 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkb46\" (UniqueName: \"kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914681 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914846 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914901 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.914930 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.915006 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.915040 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.915081 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.915120 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:36 crc kubenswrapper[4761]: I1125 23:14:36.915166 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016442 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016507 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016559 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016601 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016645 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016650 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016749 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016789 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016832 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016860 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016884 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016853 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016893 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkb46\" (UniqueName: \"kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016970 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016914 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017047 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017079 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.016853 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017099 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017131 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017153 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017273 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017279 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.017323 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.022530 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.026001 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.040426 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.041305 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkb46\" (UniqueName: \"kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.046815 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="310ccbc5-38e7-402c-8cc0-adf9c7038e42" path="/var/lib/kubelet/pods/310ccbc5-38e7-402c-8cc0-adf9c7038e42/volumes" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.046947 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.067567 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.092500 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.315170 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:14:37 crc kubenswrapper[4761]: I1125 23:14:37.662388 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerStarted","Data":"3998d980de13ee0e41b514f089c07f93cd70833ce1edb3b43f4dc7fd64e1a17a"} Nov 25 23:14:38 crc kubenswrapper[4761]: I1125 23:14:38.672869 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerStarted","Data":"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625"} Nov 25 23:14:38 crc kubenswrapper[4761]: I1125 23:14:38.673330 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerStarted","Data":"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa"} Nov 25 23:14:38 crc kubenswrapper[4761]: I1125 23:14:38.700302 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.700267626 podStartE2EDuration="2.700267626s" podCreationTimestamp="2025-11-25 23:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:14:38.699942147 +0000 UTC m=+1074.433288012" watchObservedRunningTime="2025-11-25 23:14:38.700267626 +0000 UTC m=+1074.433613511" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.092901 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.093486 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.129342 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.167643 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.754279 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:47 crc kubenswrapper[4761]: I1125 23:14:47.754341 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:49 crc kubenswrapper[4761]: I1125 23:14:49.630829 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:14:49 crc kubenswrapper[4761]: I1125 23:14:49.635997 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.163640 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5"] Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.165595 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.168432 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.168886 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.180447 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5"] Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.301923 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6hqn\" (UniqueName: \"kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.302007 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.302073 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.403491 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6hqn\" (UniqueName: \"kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.403601 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.403717 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.405276 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.409572 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.430521 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6hqn\" (UniqueName: \"kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn\") pod \"collect-profiles-29401875-qlxk5\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.502974 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.797110 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5"] Nov 25 23:15:00 crc kubenswrapper[4761]: I1125 23:15:00.868166 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" event={"ID":"30d59ea5-74ee-4393-b212-4fec0adae715","Type":"ContainerStarted","Data":"9a86fb921ab4eb1038c81492c64535f1b1849d9316e4929eb6d90add3c40f88a"} Nov 25 23:15:01 crc kubenswrapper[4761]: I1125 23:15:01.881186 4761 generic.go:334] "Generic (PLEG): container finished" podID="30d59ea5-74ee-4393-b212-4fec0adae715" containerID="625ae9d991d1901154f85357d70abcd30d5afa0c9c4dd91411f17347ba34e754" exitCode=0 Nov 25 23:15:01 crc kubenswrapper[4761]: I1125 23:15:01.881304 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" event={"ID":"30d59ea5-74ee-4393-b212-4fec0adae715","Type":"ContainerDied","Data":"625ae9d991d1901154f85357d70abcd30d5afa0c9c4dd91411f17347ba34e754"} Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.337498 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.454361 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume\") pod \"30d59ea5-74ee-4393-b212-4fec0adae715\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.454443 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6hqn\" (UniqueName: \"kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn\") pod \"30d59ea5-74ee-4393-b212-4fec0adae715\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.454499 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume\") pod \"30d59ea5-74ee-4393-b212-4fec0adae715\" (UID: \"30d59ea5-74ee-4393-b212-4fec0adae715\") " Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.456537 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume" (OuterVolumeSpecName: "config-volume") pod "30d59ea5-74ee-4393-b212-4fec0adae715" (UID: "30d59ea5-74ee-4393-b212-4fec0adae715"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.464207 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "30d59ea5-74ee-4393-b212-4fec0adae715" (UID: "30d59ea5-74ee-4393-b212-4fec0adae715"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.465161 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn" (OuterVolumeSpecName: "kube-api-access-p6hqn") pod "30d59ea5-74ee-4393-b212-4fec0adae715" (UID: "30d59ea5-74ee-4393-b212-4fec0adae715"). InnerVolumeSpecName "kube-api-access-p6hqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.556685 4761 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30d59ea5-74ee-4393-b212-4fec0adae715-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.556784 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6hqn\" (UniqueName: \"kubernetes.io/projected/30d59ea5-74ee-4393-b212-4fec0adae715-kube-api-access-p6hqn\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.556845 4761 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30d59ea5-74ee-4393-b212-4fec0adae715-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.906784 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" event={"ID":"30d59ea5-74ee-4393-b212-4fec0adae715","Type":"ContainerDied","Data":"9a86fb921ab4eb1038c81492c64535f1b1849d9316e4929eb6d90add3c40f88a"} Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.906911 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a86fb921ab4eb1038c81492c64535f1b1849d9316e4929eb6d90add3c40f88a" Nov 25 23:15:03 crc kubenswrapper[4761]: I1125 23:15:03.907006 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401875-qlxk5" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.173214 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-jw7qd"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.181822 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-jw7qd"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.264212 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.264491 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-log" containerID="cri-o://27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1" gracePeriod=30 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.264592 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-httpd" containerID="cri-o://7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855" gracePeriod=30 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.268749 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.268969 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-log" containerID="cri-o://8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa" gracePeriod=30 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.269108 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-httpd" containerID="cri-o://7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625" gracePeriod=30 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.287901 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance89b9-account-delete-wsb8p"] Nov 25 23:15:07 crc kubenswrapper[4761]: E1125 23:15:07.288163 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d59ea5-74ee-4393-b212-4fec0adae715" containerName="collect-profiles" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.288178 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d59ea5-74ee-4393-b212-4fec0adae715" containerName="collect-profiles" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.288318 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d59ea5-74ee-4393-b212-4fec0adae715" containerName="collect-profiles" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.288756 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.301535 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance89b9-account-delete-wsb8p"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.336245 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.336438 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="13642318-226b-4cf1-84fc-38d0eb5bd797" containerName="openstackclient" containerID="cri-o://73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d" gracePeriod=30 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.434899 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.435240 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4kvs\" (UniqueName: \"kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.537607 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4kvs\" (UniqueName: \"kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.537916 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.538764 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.561928 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4kvs\" (UniqueName: \"kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs\") pod \"glance89b9-account-delete-wsb8p\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.615766 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.683874 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.739779 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2cb5\" (UniqueName: \"kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5\") pod \"13642318-226b-4cf1-84fc-38d0eb5bd797\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.740219 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config\") pod \"13642318-226b-4cf1-84fc-38d0eb5bd797\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.740367 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret\") pod \"13642318-226b-4cf1-84fc-38d0eb5bd797\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.740589 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts\") pod \"13642318-226b-4cf1-84fc-38d0eb5bd797\" (UID: \"13642318-226b-4cf1-84fc-38d0eb5bd797\") " Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.751770 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "13642318-226b-4cf1-84fc-38d0eb5bd797" (UID: "13642318-226b-4cf1-84fc-38d0eb5bd797"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.754144 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5" (OuterVolumeSpecName: "kube-api-access-z2cb5") pod "13642318-226b-4cf1-84fc-38d0eb5bd797" (UID: "13642318-226b-4cf1-84fc-38d0eb5bd797"). InnerVolumeSpecName "kube-api-access-z2cb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.766654 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "13642318-226b-4cf1-84fc-38d0eb5bd797" (UID: "13642318-226b-4cf1-84fc-38d0eb5bd797"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.787141 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "13642318-226b-4cf1-84fc-38d0eb5bd797" (UID: "13642318-226b-4cf1-84fc-38d0eb5bd797"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.851188 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2cb5\" (UniqueName: \"kubernetes.io/projected/13642318-226b-4cf1-84fc-38d0eb5bd797-kube-api-access-z2cb5\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.851220 4761 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.851229 4761 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.851237 4761 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/13642318-226b-4cf1-84fc-38d0eb5bd797-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.946332 4761 generic.go:334] "Generic (PLEG): container finished" podID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerID="27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1" exitCode=143 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.946413 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerDied","Data":"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1"} Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.949204 4761 generic.go:334] "Generic (PLEG): container finished" podID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerID="8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa" exitCode=143 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.949311 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerDied","Data":"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa"} Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.953198 4761 generic.go:334] "Generic (PLEG): container finished" podID="13642318-226b-4cf1-84fc-38d0eb5bd797" containerID="73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d" exitCode=143 Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.953244 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"13642318-226b-4cf1-84fc-38d0eb5bd797","Type":"ContainerDied","Data":"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d"} Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.953273 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"13642318-226b-4cf1-84fc-38d0eb5bd797","Type":"ContainerDied","Data":"85c8d5d2c484928f99bf365502f2360d8a3ce589a37e37b971f537122c13fb6a"} Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.953300 4761 scope.go:117] "RemoveContainer" containerID="73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.953478 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.990455 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.991355 4761 scope.go:117] "RemoveContainer" containerID="73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d" Nov 25 23:15:07 crc kubenswrapper[4761]: E1125 23:15:07.992036 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d\": container with ID starting with 73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d not found: ID does not exist" containerID="73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.992073 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d"} err="failed to get container status \"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d\": rpc error: code = NotFound desc = could not find container \"73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d\": container with ID starting with 73c5d1364fe649a7dfebf3e7a759f557a77f36aec2d07f03d6d394cb56e3db9d not found: ID does not exist" Nov 25 23:15:07 crc kubenswrapper[4761]: I1125 23:15:07.996626 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:15:08 crc kubenswrapper[4761]: I1125 23:15:08.087673 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance89b9-account-delete-wsb8p"] Nov 25 23:15:08 crc kubenswrapper[4761]: I1125 23:15:08.961454 4761 generic.go:334] "Generic (PLEG): container finished" podID="e240800e-5fa7-4cbf-b2d7-4c9c68a86743" containerID="b7d0a09b081655af53b4e8ded2c2d3bb8e3539ff5beea727ee082a98215c4436" exitCode=0 Nov 25 23:15:08 crc kubenswrapper[4761]: I1125 23:15:08.961533 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" event={"ID":"e240800e-5fa7-4cbf-b2d7-4c9c68a86743","Type":"ContainerDied","Data":"b7d0a09b081655af53b4e8ded2c2d3bb8e3539ff5beea727ee082a98215c4436"} Nov 25 23:15:08 crc kubenswrapper[4761]: I1125 23:15:08.961808 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" event={"ID":"e240800e-5fa7-4cbf-b2d7-4c9c68a86743","Type":"ContainerStarted","Data":"243abc18b5d78348a10bf052eb083158e9756f525835ee828d975f449e8c5152"} Nov 25 23:15:09 crc kubenswrapper[4761]: I1125 23:15:09.022827 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13642318-226b-4cf1-84fc-38d0eb5bd797" path="/var/lib/kubelet/pods/13642318-226b-4cf1-84fc-38d0eb5bd797/volumes" Nov 25 23:15:09 crc kubenswrapper[4761]: I1125 23:15:09.023858 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d87d1f9-7c4a-4dab-bee8-6f20f4a02415" path="/var/lib/kubelet/pods/1d87d1f9-7c4a-4dab-bee8-6f20f4a02415/volumes" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.338287 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.496017 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts\") pod \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.496112 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4kvs\" (UniqueName: \"kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs\") pod \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\" (UID: \"e240800e-5fa7-4cbf-b2d7-4c9c68a86743\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.497892 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e240800e-5fa7-4cbf-b2d7-4c9c68a86743" (UID: "e240800e-5fa7-4cbf-b2d7-4c9c68a86743"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.502911 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs" (OuterVolumeSpecName: "kube-api-access-r4kvs") pod "e240800e-5fa7-4cbf-b2d7-4c9c68a86743" (UID: "e240800e-5fa7-4cbf-b2d7-4c9c68a86743"). InnerVolumeSpecName "kube-api-access-r4kvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.597447 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.597476 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4kvs\" (UniqueName: \"kubernetes.io/projected/e240800e-5fa7-4cbf-b2d7-4c9c68a86743-kube-api-access-r4kvs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.791568 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.811720 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901493 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901543 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901571 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkb46\" (UniqueName: \"kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901598 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901612 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901651 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901713 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901736 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901757 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901777 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901793 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901812 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901833 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901863 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901885 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901899 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901914 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901937 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901969 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.901988 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902009 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902026 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902052 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g87th\" (UniqueName: \"kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902064 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902081 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902105 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run\") pod \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\" (UID: \"660e9154-b8d1-468e-bbe5-bc884c4ad0a7\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902127 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902144 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data\") pod \"f56e9ad0-2818-47d8-bd16-58bf91010833\" (UID: \"f56e9ad0-2818-47d8-bd16-58bf91010833\") " Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902731 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902781 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902803 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902824 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.902843 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.903110 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.907115 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run" (OuterVolumeSpecName: "run") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.907144 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.907207 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys" (OuterVolumeSpecName: "sys") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.907216 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev" (OuterVolumeSpecName: "dev") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.908070 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.908167 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run" (OuterVolumeSpecName: "run") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.909597 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys" (OuterVolumeSpecName: "sys") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.909637 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.909920 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev" (OuterVolumeSpecName: "dev") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.909998 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts" (OuterVolumeSpecName: "scripts") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.910112 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.910217 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs" (OuterVolumeSpecName: "logs") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.910256 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts" (OuterVolumeSpecName: "scripts") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.910621 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46" (OuterVolumeSpecName: "kube-api-access-fkb46") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "kube-api-access-fkb46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.911085 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.913034 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.913094 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th" (OuterVolumeSpecName: "kube-api-access-g87th") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "kube-api-access-g87th". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.913566 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.913566 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs" (OuterVolumeSpecName: "logs") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.913760 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.953442 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data" (OuterVolumeSpecName: "config-data") pod "f56e9ad0-2818-47d8-bd16-58bf91010833" (UID: "f56e9ad0-2818-47d8-bd16-58bf91010833"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.960979 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data" (OuterVolumeSpecName: "config-data") pod "660e9154-b8d1-468e-bbe5-bc884c4ad0a7" (UID: "660e9154-b8d1-468e-bbe5-bc884c4ad0a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.981926 4761 generic.go:334] "Generic (PLEG): container finished" podID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerID="7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855" exitCode=0 Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.982014 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerDied","Data":"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855"} Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.982050 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"660e9154-b8d1-468e-bbe5-bc884c4ad0a7","Type":"ContainerDied","Data":"0379144f637967767bf0c61c3dc7469b8019fa7c1b2df88f5d993036a0e5cf16"} Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.982056 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.982072 4761 scope.go:117] "RemoveContainer" containerID="7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.984368 4761 generic.go:334] "Generic (PLEG): container finished" podID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerID="7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625" exitCode=0 Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.984400 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerDied","Data":"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625"} Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.984427 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f56e9ad0-2818-47d8-bd16-58bf91010833","Type":"ContainerDied","Data":"3998d980de13ee0e41b514f089c07f93cd70833ce1edb3b43f4dc7fd64e1a17a"} Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.984459 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.988955 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" event={"ID":"e240800e-5fa7-4cbf-b2d7-4c9c68a86743","Type":"ContainerDied","Data":"243abc18b5d78348a10bf052eb083158e9756f525835ee828d975f449e8c5152"} Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.989121 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="243abc18b5d78348a10bf052eb083158e9756f525835ee828d975f449e8c5152" Nov 25 23:15:10 crc kubenswrapper[4761]: I1125 23:15:10.989271 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance89b9-account-delete-wsb8p" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004160 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004204 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004221 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004237 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004253 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g87th\" (UniqueName: \"kubernetes.io/projected/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-kube-api-access-g87th\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004307 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004332 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004349 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004366 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004381 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e9ad0-2818-47d8-bd16-58bf91010833-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004397 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004411 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004426 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkb46\" (UniqueName: \"kubernetes.io/projected/f56e9ad0-2818-47d8-bd16-58bf91010833-kube-api-access-fkb46\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004440 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004455 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004470 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e9ad0-2818-47d8-bd16-58bf91010833-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004486 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004502 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004519 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004535 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004550 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004566 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004582 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004597 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004635 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004660 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004678 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f56e9ad0-2818-47d8-bd16-58bf91010833-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.004722 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/660e9154-b8d1-468e-bbe5-bc884c4ad0a7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.032445 4761 scope.go:117] "RemoveContainer" containerID="27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.052934 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.053057 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.068559 4761 scope.go:117] "RemoveContainer" containerID="7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855" Nov 25 23:15:11 crc kubenswrapper[4761]: E1125 23:15:11.069369 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855\": container with ID starting with 7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855 not found: ID does not exist" containerID="7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.069403 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855"} err="failed to get container status \"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855\": rpc error: code = NotFound desc = could not find container \"7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855\": container with ID starting with 7e970a09204176ed2e422b2319dd6168f8233ea9e4a055634673ff1663ec3855 not found: ID does not exist" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.069426 4761 scope.go:117] "RemoveContainer" containerID="27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.069477 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 23:15:11 crc kubenswrapper[4761]: E1125 23:15:11.071509 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1\": container with ID starting with 27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1 not found: ID does not exist" containerID="27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.071665 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1"} err="failed to get container status \"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1\": rpc error: code = NotFound desc = could not find container \"27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1\": container with ID starting with 27b3d887f01c274ef68b75a6e9357e393b089cec34a90d15d418c340207aafe1 not found: ID does not exist" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.071815 4761 scope.go:117] "RemoveContainer" containerID="7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.076348 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.077649 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.077684 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.084162 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.096879 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.100762 4761 scope.go:117] "RemoveContainer" containerID="8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.105991 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.106018 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.106026 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.106034 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.120121 4761 scope.go:117] "RemoveContainer" containerID="7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625" Nov 25 23:15:11 crc kubenswrapper[4761]: E1125 23:15:11.120781 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625\": container with ID starting with 7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625 not found: ID does not exist" containerID="7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.120836 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625"} err="failed to get container status \"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625\": rpc error: code = NotFound desc = could not find container \"7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625\": container with ID starting with 7926bc9a44f8a553e4a09d4156a528be799547d1704e8f9e911353341f99d625 not found: ID does not exist" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.120878 4761 scope.go:117] "RemoveContainer" containerID="8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa" Nov 25 23:15:11 crc kubenswrapper[4761]: E1125 23:15:11.121422 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa\": container with ID starting with 8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa not found: ID does not exist" containerID="8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa" Nov 25 23:15:11 crc kubenswrapper[4761]: I1125 23:15:11.121479 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa"} err="failed to get container status \"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa\": rpc error: code = NotFound desc = could not find container \"8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa\": container with ID starting with 8dfd4ac21e812b40d0b635d2f89415623cb8149d52afb35b47488fdd6af536fa not found: ID does not exist" Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.320663 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-kv6cr"] Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.331574 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-kv6cr"] Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.347379 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-89b9-account-create-update-nk6dv"] Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.357027 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance89b9-account-delete-wsb8p"] Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.365402 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance89b9-account-delete-wsb8p"] Nov 25 23:15:12 crc kubenswrapper[4761]: I1125 23:15:12.370981 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-89b9-account-create-update-nk6dv"] Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.026798 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a06b3e4-da6e-40a5-bec8-3fcd8aefceae" path="/var/lib/kubelet/pods/3a06b3e4-da6e-40a5-bec8-3fcd8aefceae/volumes" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.028146 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" path="/var/lib/kubelet/pods/660e9154-b8d1-468e-bbe5-bc884c4ad0a7/volumes" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.029244 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="791ea0cf-ae15-4e32-911d-93c65978db8d" path="/var/lib/kubelet/pods/791ea0cf-ae15-4e32-911d-93c65978db8d/volumes" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.031101 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e240800e-5fa7-4cbf-b2d7-4c9c68a86743" path="/var/lib/kubelet/pods/e240800e-5fa7-4cbf-b2d7-4c9c68a86743/volumes" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.032186 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" path="/var/lib/kubelet/pods/f56e9ad0-2818-47d8-bd16-58bf91010833/volumes" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.400596 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-k52nj"] Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401019 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e240800e-5fa7-4cbf-b2d7-4c9c68a86743" containerName="mariadb-account-delete" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401040 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="e240800e-5fa7-4cbf-b2d7-4c9c68a86743" containerName="mariadb-account-delete" Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401072 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13642318-226b-4cf1-84fc-38d0eb5bd797" containerName="openstackclient" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401084 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="13642318-226b-4cf1-84fc-38d0eb5bd797" containerName="openstackclient" Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401107 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401120 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401143 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401155 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401174 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401186 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: E1125 23:15:13.401204 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401216 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401452 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="13642318-226b-4cf1-84fc-38d0eb5bd797" containerName="openstackclient" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401476 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401505 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="e240800e-5fa7-4cbf-b2d7-4c9c68a86743" containerName="mariadb-account-delete" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401525 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-log" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401542 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="660e9154-b8d1-468e-bbe5-bc884c4ad0a7" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.401561 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56e9ad0-2818-47d8-bd16-58bf91010833" containerName="glance-httpd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.402321 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.414341 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-455d-account-create-update-5xgtd"] Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.415520 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.418717 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.424107 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-k52nj"] Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.437005 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-455d-account-create-update-5xgtd"] Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.544147 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4ngd\" (UniqueName: \"kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.544503 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.544625 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xzmr\" (UniqueName: \"kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.544829 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.646369 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4ngd\" (UniqueName: \"kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.646442 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.646466 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xzmr\" (UniqueName: \"kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.646560 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.647363 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.647397 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.675879 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xzmr\" (UniqueName: \"kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr\") pod \"glance-db-create-k52nj\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.677205 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4ngd\" (UniqueName: \"kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd\") pod \"glance-455d-account-create-update-5xgtd\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.723381 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:13 crc kubenswrapper[4761]: I1125 23:15:13.744236 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:14 crc kubenswrapper[4761]: I1125 23:15:14.162658 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-k52nj"] Nov 25 23:15:14 crc kubenswrapper[4761]: I1125 23:15:14.202279 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-455d-account-create-update-5xgtd"] Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.033130 4761 generic.go:334] "Generic (PLEG): container finished" podID="4575c493-2259-4ba8-bd9f-e5bc51abd0bf" containerID="74b2da04d0c1f7778e6fcbc66f71cef9e69b598da0ef2ee653d17a1e389105fd" exitCode=0 Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.033245 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" event={"ID":"4575c493-2259-4ba8-bd9f-e5bc51abd0bf","Type":"ContainerDied","Data":"74b2da04d0c1f7778e6fcbc66f71cef9e69b598da0ef2ee653d17a1e389105fd"} Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.033291 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" event={"ID":"4575c493-2259-4ba8-bd9f-e5bc51abd0bf","Type":"ContainerStarted","Data":"06e8caa026535c6ef227f6dd45fdad111ea07724b4b899cf80f6ca51dd8a6211"} Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.036008 4761 generic.go:334] "Generic (PLEG): container finished" podID="dd597ebf-c55b-4905-989c-8e9185ed579f" containerID="a3541be67296fdb6f307d37a883e8d6326768c9d2d067c1fd4fa009f7a7acf31" exitCode=0 Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.036060 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k52nj" event={"ID":"dd597ebf-c55b-4905-989c-8e9185ed579f","Type":"ContainerDied","Data":"a3541be67296fdb6f307d37a883e8d6326768c9d2d067c1fd4fa009f7a7acf31"} Nov 25 23:15:15 crc kubenswrapper[4761]: I1125 23:15:15.036116 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k52nj" event={"ID":"dd597ebf-c55b-4905-989c-8e9185ed579f","Type":"ContainerStarted","Data":"3bae0c997e6b4d2f89513a1fa2970443d87d00d1eb417052783e98a351f83df6"} Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.342526 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.390722 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts\") pod \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.390829 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4ngd\" (UniqueName: \"kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd\") pod \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\" (UID: \"4575c493-2259-4ba8-bd9f-e5bc51abd0bf\") " Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.391793 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4575c493-2259-4ba8-bd9f-e5bc51abd0bf" (UID: "4575c493-2259-4ba8-bd9f-e5bc51abd0bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.399154 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd" (OuterVolumeSpecName: "kube-api-access-m4ngd") pod "4575c493-2259-4ba8-bd9f-e5bc51abd0bf" (UID: "4575c493-2259-4ba8-bd9f-e5bc51abd0bf"). InnerVolumeSpecName "kube-api-access-m4ngd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.436191 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.491988 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xzmr\" (UniqueName: \"kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr\") pod \"dd597ebf-c55b-4905-989c-8e9185ed579f\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.492162 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts\") pod \"dd597ebf-c55b-4905-989c-8e9185ed579f\" (UID: \"dd597ebf-c55b-4905-989c-8e9185ed579f\") " Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.492791 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.492818 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4ngd\" (UniqueName: \"kubernetes.io/projected/4575c493-2259-4ba8-bd9f-e5bc51abd0bf-kube-api-access-m4ngd\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.493015 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd597ebf-c55b-4905-989c-8e9185ed579f" (UID: "dd597ebf-c55b-4905-989c-8e9185ed579f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.496183 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr" (OuterVolumeSpecName: "kube-api-access-5xzmr") pod "dd597ebf-c55b-4905-989c-8e9185ed579f" (UID: "dd597ebf-c55b-4905-989c-8e9185ed579f"). InnerVolumeSpecName "kube-api-access-5xzmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.594254 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd597ebf-c55b-4905-989c-8e9185ed579f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:16 crc kubenswrapper[4761]: I1125 23:15:16.594294 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xzmr\" (UniqueName: \"kubernetes.io/projected/dd597ebf-c55b-4905-989c-8e9185ed579f-kube-api-access-5xzmr\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.051388 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" event={"ID":"4575c493-2259-4ba8-bd9f-e5bc51abd0bf","Type":"ContainerDied","Data":"06e8caa026535c6ef227f6dd45fdad111ea07724b4b899cf80f6ca51dd8a6211"} Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.051444 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06e8caa026535c6ef227f6dd45fdad111ea07724b4b899cf80f6ca51dd8a6211" Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.051426 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-455d-account-create-update-5xgtd" Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.053493 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k52nj" event={"ID":"dd597ebf-c55b-4905-989c-8e9185ed579f","Type":"ContainerDied","Data":"3bae0c997e6b4d2f89513a1fa2970443d87d00d1eb417052783e98a351f83df6"} Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.053565 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bae0c997e6b4d2f89513a1fa2970443d87d00d1eb417052783e98a351f83df6" Nov 25 23:15:17 crc kubenswrapper[4761]: I1125 23:15:17.053542 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k52nj" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.723109 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-x5zjt"] Nov 25 23:15:18 crc kubenswrapper[4761]: E1125 23:15:18.723572 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd597ebf-c55b-4905-989c-8e9185ed579f" containerName="mariadb-database-create" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.723584 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd597ebf-c55b-4905-989c-8e9185ed579f" containerName="mariadb-database-create" Nov 25 23:15:18 crc kubenswrapper[4761]: E1125 23:15:18.723608 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4575c493-2259-4ba8-bd9f-e5bc51abd0bf" containerName="mariadb-account-create-update" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.723614 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="4575c493-2259-4ba8-bd9f-e5bc51abd0bf" containerName="mariadb-account-create-update" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.723742 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="4575c493-2259-4ba8-bd9f-e5bc51abd0bf" containerName="mariadb-account-create-update" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.723754 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd597ebf-c55b-4905-989c-8e9185ed579f" containerName="mariadb-database-create" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.724186 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.726256 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.726740 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qs6gz" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.735383 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x5zjt"] Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.737837 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.842652 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.842687 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5jsx\" (UniqueName: \"kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.842934 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.843003 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.945102 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5jsx\" (UniqueName: \"kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.945174 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.945288 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.945329 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.949649 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.950078 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.963753 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:18 crc kubenswrapper[4761]: I1125 23:15:18.984423 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5jsx\" (UniqueName: \"kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx\") pod \"glance-db-sync-x5zjt\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:19 crc kubenswrapper[4761]: I1125 23:15:19.050769 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:19 crc kubenswrapper[4761]: I1125 23:15:19.545017 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x5zjt"] Nov 25 23:15:20 crc kubenswrapper[4761]: I1125 23:15:20.079426 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x5zjt" event={"ID":"ed20e705-51cc-434d-a4dc-92eaa5ee38e3","Type":"ContainerStarted","Data":"4ea99e0bff7bd714c7e79423ae3f83dcdd577ea93912f9d52703a5c47c12e786"} Nov 25 23:15:21 crc kubenswrapper[4761]: I1125 23:15:21.088926 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x5zjt" event={"ID":"ed20e705-51cc-434d-a4dc-92eaa5ee38e3","Type":"ContainerStarted","Data":"8d08c0aa3234212faeeee9a61a65a81bef37e77cdfc35ae022b7b64b50213129"} Nov 25 23:15:21 crc kubenswrapper[4761]: I1125 23:15:21.112838 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-x5zjt" podStartSLOduration=3.112813181 podStartE2EDuration="3.112813181s" podCreationTimestamp="2025-11-25 23:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:15:21.106027971 +0000 UTC m=+1116.839373846" watchObservedRunningTime="2025-11-25 23:15:21.112813181 +0000 UTC m=+1116.846159026" Nov 25 23:15:21 crc kubenswrapper[4761]: I1125 23:15:21.127456 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:15:21 crc kubenswrapper[4761]: I1125 23:15:21.127528 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:15:23 crc kubenswrapper[4761]: I1125 23:15:23.112309 4761 generic.go:334] "Generic (PLEG): container finished" podID="ed20e705-51cc-434d-a4dc-92eaa5ee38e3" containerID="8d08c0aa3234212faeeee9a61a65a81bef37e77cdfc35ae022b7b64b50213129" exitCode=0 Nov 25 23:15:23 crc kubenswrapper[4761]: I1125 23:15:23.112384 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x5zjt" event={"ID":"ed20e705-51cc-434d-a4dc-92eaa5ee38e3","Type":"ContainerDied","Data":"8d08c0aa3234212faeeee9a61a65a81bef37e77cdfc35ae022b7b64b50213129"} Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.501123 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.694329 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data\") pod \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.694498 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5jsx\" (UniqueName: \"kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx\") pod \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.694622 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data\") pod \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.694691 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle\") pod \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\" (UID: \"ed20e705-51cc-434d-a4dc-92eaa5ee38e3\") " Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.702935 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ed20e705-51cc-434d-a4dc-92eaa5ee38e3" (UID: "ed20e705-51cc-434d-a4dc-92eaa5ee38e3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.703510 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx" (OuterVolumeSpecName: "kube-api-access-j5jsx") pod "ed20e705-51cc-434d-a4dc-92eaa5ee38e3" (UID: "ed20e705-51cc-434d-a4dc-92eaa5ee38e3"). InnerVolumeSpecName "kube-api-access-j5jsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.752753 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed20e705-51cc-434d-a4dc-92eaa5ee38e3" (UID: "ed20e705-51cc-434d-a4dc-92eaa5ee38e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.783125 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data" (OuterVolumeSpecName: "config-data") pod "ed20e705-51cc-434d-a4dc-92eaa5ee38e3" (UID: "ed20e705-51cc-434d-a4dc-92eaa5ee38e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.797018 4761 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.797052 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5jsx\" (UniqueName: \"kubernetes.io/projected/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-kube-api-access-j5jsx\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.797065 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:24 crc kubenswrapper[4761]: I1125 23:15:24.797079 4761 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed20e705-51cc-434d-a4dc-92eaa5ee38e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:25 crc kubenswrapper[4761]: I1125 23:15:25.131279 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x5zjt" event={"ID":"ed20e705-51cc-434d-a4dc-92eaa5ee38e3","Type":"ContainerDied","Data":"4ea99e0bff7bd714c7e79423ae3f83dcdd577ea93912f9d52703a5c47c12e786"} Nov 25 23:15:25 crc kubenswrapper[4761]: I1125 23:15:25.131323 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ea99e0bff7bd714c7e79423ae3f83dcdd577ea93912f9d52703a5c47c12e786" Nov 25 23:15:25 crc kubenswrapper[4761]: I1125 23:15:25.131388 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x5zjt" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.555734 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:26 crc kubenswrapper[4761]: E1125 23:15:26.556197 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed20e705-51cc-434d-a4dc-92eaa5ee38e3" containerName="glance-db-sync" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.556211 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed20e705-51cc-434d-a4dc-92eaa5ee38e3" containerName="glance-db-sync" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.556337 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed20e705-51cc-434d-a4dc-92eaa5ee38e3" containerName="glance-db-sync" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.557008 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.559042 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.559642 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.559687 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.561922 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.562427 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.563373 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qs6gz" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.573559 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.619833 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:26 crc kubenswrapper[4761]: E1125 23:15:26.620339 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-ncb98 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-ncb98 logs public-tls-certs scripts]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="14700942-139a-4988-9334-1c2d814c681b" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635214 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635270 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635289 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635313 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635344 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635364 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncb98\" (UniqueName: \"kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635380 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635421 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.635435 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.736862 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737088 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737182 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737249 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737334 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737399 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737459 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737532 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737617 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncb98\" (UniqueName: \"kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737759 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737770 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.737929 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.743208 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.743215 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.743480 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.743521 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.744252 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.757635 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncb98\" (UniqueName: \"kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:26 crc kubenswrapper[4761]: I1125 23:15:26.759103 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.149194 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.164071 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.241265 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.241885 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242092 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242264 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242457 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242619 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242801 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncb98\" (UniqueName: \"kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.242956 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.243087 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"14700942-139a-4988-9334-1c2d814c681b\" (UID: \"14700942-139a-4988-9334-1c2d814c681b\") " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.243717 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.243788 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs" (OuterVolumeSpecName: "logs") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.248199 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data" (OuterVolumeSpecName: "config-data") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.248246 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.248927 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.249511 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98" (OuterVolumeSpecName: "kube-api-access-ncb98") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "kube-api-access-ncb98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.249924 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.250536 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.252151 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts" (OuterVolumeSpecName: "scripts") pod "14700942-139a-4988-9334-1c2d814c681b" (UID: "14700942-139a-4988-9334-1c2d814c681b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345727 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345777 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345797 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345816 4761 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345834 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14700942-139a-4988-9334-1c2d814c681b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345850 4761 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345879 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncb98\" (UniqueName: \"kubernetes.io/projected/14700942-139a-4988-9334-1c2d814c681b-kube-api-access-ncb98\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.345979 4761 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14700942-139a-4988-9334-1c2d814c681b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.346035 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.373890 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 23:15:27 crc kubenswrapper[4761]: I1125 23:15:27.448315 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.158037 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.244629 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.263805 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.275236 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.276866 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.279473 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.279660 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.279802 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.280125 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qs6gz" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.280327 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.280817 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.298878 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464416 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464513 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464547 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464605 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464643 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.464925 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8tf2\" (UniqueName: \"kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.465015 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.465096 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.465145 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567113 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567234 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567277 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567343 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567388 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567397 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567455 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8tf2\" (UniqueName: \"kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567595 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567662 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.567792 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.568341 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.569145 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.574860 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.574943 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.576645 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.577196 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.583925 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.599955 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.606769 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8tf2\" (UniqueName: \"kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2\") pod \"glance-default-single-0\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:28 crc kubenswrapper[4761]: I1125 23:15:28.907068 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:29 crc kubenswrapper[4761]: I1125 23:15:29.031466 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14700942-139a-4988-9334-1c2d814c681b" path="/var/lib/kubelet/pods/14700942-139a-4988-9334-1c2d814c681b/volumes" Nov 25 23:15:29 crc kubenswrapper[4761]: I1125 23:15:29.373430 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:30 crc kubenswrapper[4761]: I1125 23:15:30.179103 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerStarted","Data":"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5"} Nov 25 23:15:30 crc kubenswrapper[4761]: I1125 23:15:30.179414 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerStarted","Data":"8e78d8dbb32a9f7140de250d42217d403183422ae689cbfc7c4a9a1f1cae2c7e"} Nov 25 23:15:31 crc kubenswrapper[4761]: I1125 23:15:31.194397 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerStarted","Data":"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96"} Nov 25 23:15:31 crc kubenswrapper[4761]: I1125 23:15:31.231996 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.231969947 podStartE2EDuration="3.231969947s" podCreationTimestamp="2025-11-25 23:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:15:31.227278503 +0000 UTC m=+1126.960624418" watchObservedRunningTime="2025-11-25 23:15:31.231969947 +0000 UTC m=+1126.965315822" Nov 25 23:15:38 crc kubenswrapper[4761]: I1125 23:15:38.907952 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:38 crc kubenswrapper[4761]: I1125 23:15:38.908651 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:38 crc kubenswrapper[4761]: I1125 23:15:38.960889 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:38 crc kubenswrapper[4761]: I1125 23:15:38.979178 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:39 crc kubenswrapper[4761]: I1125 23:15:39.269538 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:39 crc kubenswrapper[4761]: I1125 23:15:39.269899 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:41 crc kubenswrapper[4761]: I1125 23:15:41.188978 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:41 crc kubenswrapper[4761]: I1125 23:15:41.221645 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.152938 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x5zjt"] Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.164027 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x5zjt"] Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.225760 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance455d-account-delete-hvp2n"] Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.226726 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.230761 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance455d-account-delete-hvp2n"] Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.250752 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.292818 4761 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/glance-default-single-0" secret="" err="secret \"glance-glance-dockercfg-qs6gz\" not found" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.293725 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m68m\" (UniqueName: \"kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.293843 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.395384 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m68m\" (UniqueName: \"kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.395427 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.395960 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.395987 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.396037 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:42.896017787 +0000 UTC m=+1138.629363622 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-default-single-config-data" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.396056 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:42.896049338 +0000 UTC m=+1138.629395173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-scripts" not found Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.396605 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.413547 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m68m\" (UniqueName: \"kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m\") pod \"glance455d-account-delete-hvp2n\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: I1125 23:15:42.557541 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.902982 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.902991 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.903299 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:43.903279758 +0000 UTC m=+1139.636625603 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-default-single-config-data" not found Nov 25 23:15:42 crc kubenswrapper[4761]: E1125 23:15:42.903366 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:43.903331089 +0000 UTC m=+1139.636676934 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-scripts" not found Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.021492 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed20e705-51cc-434d-a4dc-92eaa5ee38e3" path="/var/lib/kubelet/pods/ed20e705-51cc-434d-a4dc-92eaa5ee38e3/volumes" Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.022453 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance455d-account-delete-hvp2n"] Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.301049 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-log" containerID="cri-o://d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5" gracePeriod=30 Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.302329 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" event={"ID":"909d46d6-91f2-4009-9afe-5fe2a9e362ef","Type":"ContainerStarted","Data":"f4d902e8a57f9b519513f37a8f3fa687f603d8cb2a868c355c857cc73dae40c0"} Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.302371 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" event={"ID":"909d46d6-91f2-4009-9afe-5fe2a9e362ef","Type":"ContainerStarted","Data":"9b88b0b02e41fc7302fcdd2355dcc16f00febbe4eced3d173f5138f36ebc4a4f"} Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.302655 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-httpd" containerID="cri-o://4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96" gracePeriod=30 Nov 25 23:15:43 crc kubenswrapper[4761]: I1125 23:15:43.326209 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" podStartSLOduration=1.326189418 podStartE2EDuration="1.326189418s" podCreationTimestamp="2025-11-25 23:15:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:15:43.32138626 +0000 UTC m=+1139.054732105" watchObservedRunningTime="2025-11-25 23:15:43.326189418 +0000 UTC m=+1139.059535273" Nov 25 23:15:43 crc kubenswrapper[4761]: E1125 23:15:43.939860 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 23:15:43 crc kubenswrapper[4761]: E1125 23:15:43.940263 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:45.940245664 +0000 UTC m=+1141.673591499 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-scripts" not found Nov 25 23:15:43 crc kubenswrapper[4761]: E1125 23:15:43.939877 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 23:15:43 crc kubenswrapper[4761]: E1125 23:15:43.940413 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:45.940385749 +0000 UTC m=+1141.673731634 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-default-single-config-data" not found Nov 25 23:15:44 crc kubenswrapper[4761]: I1125 23:15:44.313931 4761 generic.go:334] "Generic (PLEG): container finished" podID="909d46d6-91f2-4009-9afe-5fe2a9e362ef" containerID="f4d902e8a57f9b519513f37a8f3fa687f603d8cb2a868c355c857cc73dae40c0" exitCode=0 Nov 25 23:15:44 crc kubenswrapper[4761]: I1125 23:15:44.314056 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" event={"ID":"909d46d6-91f2-4009-9afe-5fe2a9e362ef","Type":"ContainerDied","Data":"f4d902e8a57f9b519513f37a8f3fa687f603d8cb2a868c355c857cc73dae40c0"} Nov 25 23:15:44 crc kubenswrapper[4761]: I1125 23:15:44.316910 4761 generic.go:334] "Generic (PLEG): container finished" podID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerID="d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5" exitCode=143 Nov 25 23:15:44 crc kubenswrapper[4761]: I1125 23:15:44.316956 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerDied","Data":"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5"} Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.682477 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.768646 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m68m\" (UniqueName: \"kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m\") pod \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.768762 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts\") pod \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\" (UID: \"909d46d6-91f2-4009-9afe-5fe2a9e362ef\") " Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.769308 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "909d46d6-91f2-4009-9afe-5fe2a9e362ef" (UID: "909d46d6-91f2-4009-9afe-5fe2a9e362ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.790513 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m" (OuterVolumeSpecName: "kube-api-access-6m68m") pod "909d46d6-91f2-4009-9afe-5fe2a9e362ef" (UID: "909d46d6-91f2-4009-9afe-5fe2a9e362ef"). InnerVolumeSpecName "kube-api-access-6m68m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.870404 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/909d46d6-91f2-4009-9afe-5fe2a9e362ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:45 crc kubenswrapper[4761]: I1125 23:15:45.870437 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m68m\" (UniqueName: \"kubernetes.io/projected/909d46d6-91f2-4009-9afe-5fe2a9e362ef-kube-api-access-6m68m\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:45 crc kubenswrapper[4761]: E1125 23:15:45.972618 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 23:15:45 crc kubenswrapper[4761]: E1125 23:15:45.972723 4761 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 23:15:45 crc kubenswrapper[4761]: E1125 23:15:45.972752 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:49.972720587 +0000 UTC m=+1145.706066452 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-scripts" not found Nov 25 23:15:45 crc kubenswrapper[4761]: E1125 23:15:45.972844 4761 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data podName:533c8a97-de1a-40ed-a5ed-8fb969079496 nodeName:}" failed. No retries permitted until 2025-11-25 23:15:49.972817019 +0000 UTC m=+1145.706162894 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data") pod "glance-default-single-0" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496") : secret "glance-default-single-config-data" not found Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.350264 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" event={"ID":"909d46d6-91f2-4009-9afe-5fe2a9e362ef","Type":"ContainerDied","Data":"9b88b0b02e41fc7302fcdd2355dcc16f00febbe4eced3d173f5138f36ebc4a4f"} Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.350325 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b88b0b02e41fc7302fcdd2355dcc16f00febbe4eced3d173f5138f36ebc4a4f" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.350322 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance455d-account-delete-hvp2n" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.912125 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988169 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988277 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988303 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8tf2\" (UniqueName: \"kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988325 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988358 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988377 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988411 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988479 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.988504 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"533c8a97-de1a-40ed-a5ed-8fb969079496\" (UID: \"533c8a97-de1a-40ed-a5ed-8fb969079496\") " Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.991978 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.992506 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs" (OuterVolumeSpecName: "logs") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.996365 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:15:46 crc kubenswrapper[4761]: I1125 23:15:46.997067 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts" (OuterVolumeSpecName: "scripts") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.006168 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2" (OuterVolumeSpecName: "kube-api-access-x8tf2") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "kube-api-access-x8tf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.026225 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data" (OuterVolumeSpecName: "config-data") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.030095 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.048239 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.058301 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "533c8a97-de1a-40ed-a5ed-8fb969079496" (UID: "533c8a97-de1a-40ed-a5ed-8fb969079496"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090024 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8tf2\" (UniqueName: \"kubernetes.io/projected/533c8a97-de1a-40ed-a5ed-8fb969079496-kube-api-access-x8tf2\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090058 4761 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090067 4761 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090075 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090083 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090093 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090110 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090119 4761 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533c8a97-de1a-40ed-a5ed-8fb969079496-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.090128 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533c8a97-de1a-40ed-a5ed-8fb969079496-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.102577 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.191919 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.236690 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-k52nj"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.244028 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-k52nj"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.251055 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance455d-account-delete-hvp2n"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.258305 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-455d-account-create-update-5xgtd"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.264451 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance455d-account-delete-hvp2n"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.270349 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-455d-account-create-update-5xgtd"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.362650 4761 generic.go:334] "Generic (PLEG): container finished" podID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerID="4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96" exitCode=0 Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.362767 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerDied","Data":"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96"} Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.362847 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"533c8a97-de1a-40ed-a5ed-8fb969079496","Type":"ContainerDied","Data":"8e78d8dbb32a9f7140de250d42217d403183422ae689cbfc7c4a9a1f1cae2c7e"} Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.362889 4761 scope.go:117] "RemoveContainer" containerID="4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.363302 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.391833 4761 scope.go:117] "RemoveContainer" containerID="d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.412720 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.419836 4761 scope.go:117] "RemoveContainer" containerID="4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.420020 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 23:15:47 crc kubenswrapper[4761]: E1125 23:15:47.421243 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96\": container with ID starting with 4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96 not found: ID does not exist" containerID="4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.421315 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96"} err="failed to get container status \"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96\": rpc error: code = NotFound desc = could not find container \"4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96\": container with ID starting with 4b4e7ecdb02320b2ba2bdc15b8367b8c8165f2fd6527a6058a33877f8c2f4f96 not found: ID does not exist" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.421349 4761 scope.go:117] "RemoveContainer" containerID="d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5" Nov 25 23:15:47 crc kubenswrapper[4761]: E1125 23:15:47.421848 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5\": container with ID starting with d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5 not found: ID does not exist" containerID="d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5" Nov 25 23:15:47 crc kubenswrapper[4761]: I1125 23:15:47.421886 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5"} err="failed to get container status \"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5\": rpc error: code = NotFound desc = could not find container \"d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5\": container with ID starting with d9ed094378083356d11a58d046895dacebc2dd71e76f608c6ac3b19751a205d5 not found: ID does not exist" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.494751 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-j2kwl"] Nov 25 23:15:48 crc kubenswrapper[4761]: E1125 23:15:48.495546 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-log" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.495570 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-log" Nov 25 23:15:48 crc kubenswrapper[4761]: E1125 23:15:48.495593 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909d46d6-91f2-4009-9afe-5fe2a9e362ef" containerName="mariadb-account-delete" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.495607 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="909d46d6-91f2-4009-9afe-5fe2a9e362ef" containerName="mariadb-account-delete" Nov 25 23:15:48 crc kubenswrapper[4761]: E1125 23:15:48.495677 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-httpd" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.495692 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-httpd" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.495961 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-httpd" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.496004 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="909d46d6-91f2-4009-9afe-5fe2a9e362ef" containerName="mariadb-account-delete" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.496045 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" containerName="glance-log" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.496779 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.506138 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-32e9-account-create-update-f8gtl"] Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.507898 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.516278 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.528978 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-j2kwl"] Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.553153 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-32e9-account-create-update-f8gtl"] Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.613064 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.613249 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vkk\" (UniqueName: \"kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.613453 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.614014 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjm8d\" (UniqueName: \"kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.715633 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.715707 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjm8d\" (UniqueName: \"kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.715747 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.715798 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vkk\" (UniqueName: \"kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.716604 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.716627 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.731444 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vkk\" (UniqueName: \"kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk\") pod \"glance-db-create-j2kwl\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.732155 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjm8d\" (UniqueName: \"kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d\") pod \"glance-32e9-account-create-update-f8gtl\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.832023 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:48 crc kubenswrapper[4761]: I1125 23:15:48.844630 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.021856 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4575c493-2259-4ba8-bd9f-e5bc51abd0bf" path="/var/lib/kubelet/pods/4575c493-2259-4ba8-bd9f-e5bc51abd0bf/volumes" Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.022957 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="533c8a97-de1a-40ed-a5ed-8fb969079496" path="/var/lib/kubelet/pods/533c8a97-de1a-40ed-a5ed-8fb969079496/volumes" Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.023835 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="909d46d6-91f2-4009-9afe-5fe2a9e362ef" path="/var/lib/kubelet/pods/909d46d6-91f2-4009-9afe-5fe2a9e362ef/volumes" Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.025028 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd597ebf-c55b-4905-989c-8e9185ed579f" path="/var/lib/kubelet/pods/dd597ebf-c55b-4905-989c-8e9185ed579f/volumes" Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.335609 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-j2kwl"] Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.385761 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-j2kwl" event={"ID":"01193382-e383-4185-bbd6-af046b634069","Type":"ContainerStarted","Data":"f22546f09145bdf9f35ed211a8538e7e2cec792835c80bea7c5159dd79071c5d"} Nov 25 23:15:49 crc kubenswrapper[4761]: I1125 23:15:49.406516 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-32e9-account-create-update-f8gtl"] Nov 25 23:15:49 crc kubenswrapper[4761]: W1125 23:15:49.420395 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f11741e_da8f_4bb7_abf3_038efc7ee2a6.slice/crio-f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511 WatchSource:0}: Error finding container f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511: Status 404 returned error can't find the container with id f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511 Nov 25 23:15:50 crc kubenswrapper[4761]: I1125 23:15:50.399103 4761 generic.go:334] "Generic (PLEG): container finished" podID="01193382-e383-4185-bbd6-af046b634069" containerID="2dee78dea989d0c0045e9acf721be56a08f8dc0cb5b3273457d25476cc34a728" exitCode=0 Nov 25 23:15:50 crc kubenswrapper[4761]: I1125 23:15:50.399243 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-j2kwl" event={"ID":"01193382-e383-4185-bbd6-af046b634069","Type":"ContainerDied","Data":"2dee78dea989d0c0045e9acf721be56a08f8dc0cb5b3273457d25476cc34a728"} Nov 25 23:15:50 crc kubenswrapper[4761]: I1125 23:15:50.402023 4761 generic.go:334] "Generic (PLEG): container finished" podID="4f11741e-da8f-4bb7-abf3-038efc7ee2a6" containerID="998f4d4c23f0f09f8662f1698cd2759310fd9c740033f4cd870ffb534d76e1f9" exitCode=0 Nov 25 23:15:50 crc kubenswrapper[4761]: I1125 23:15:50.402086 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" event={"ID":"4f11741e-da8f-4bb7-abf3-038efc7ee2a6","Type":"ContainerDied","Data":"998f4d4c23f0f09f8662f1698cd2759310fd9c740033f4cd870ffb534d76e1f9"} Nov 25 23:15:50 crc kubenswrapper[4761]: I1125 23:15:50.402126 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" event={"ID":"4f11741e-da8f-4bb7-abf3-038efc7ee2a6","Type":"ContainerStarted","Data":"f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511"} Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.127992 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.128520 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.865179 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.871005 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.965576 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjm8d\" (UniqueName: \"kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d\") pod \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.965654 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts\") pod \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\" (UID: \"4f11741e-da8f-4bb7-abf3-038efc7ee2a6\") " Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.965738 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts\") pod \"01193382-e383-4185-bbd6-af046b634069\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.965761 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9vkk\" (UniqueName: \"kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk\") pod \"01193382-e383-4185-bbd6-af046b634069\" (UID: \"01193382-e383-4185-bbd6-af046b634069\") " Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.966488 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01193382-e383-4185-bbd6-af046b634069" (UID: "01193382-e383-4185-bbd6-af046b634069"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.966505 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f11741e-da8f-4bb7-abf3-038efc7ee2a6" (UID: "4f11741e-da8f-4bb7-abf3-038efc7ee2a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.971241 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d" (OuterVolumeSpecName: "kube-api-access-bjm8d") pod "4f11741e-da8f-4bb7-abf3-038efc7ee2a6" (UID: "4f11741e-da8f-4bb7-abf3-038efc7ee2a6"). InnerVolumeSpecName "kube-api-access-bjm8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:51 crc kubenswrapper[4761]: I1125 23:15:51.971396 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk" (OuterVolumeSpecName: "kube-api-access-t9vkk") pod "01193382-e383-4185-bbd6-af046b634069" (UID: "01193382-e383-4185-bbd6-af046b634069"). InnerVolumeSpecName "kube-api-access-t9vkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.067240 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjm8d\" (UniqueName: \"kubernetes.io/projected/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-kube-api-access-bjm8d\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.067276 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f11741e-da8f-4bb7-abf3-038efc7ee2a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.067287 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01193382-e383-4185-bbd6-af046b634069-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.067296 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9vkk\" (UniqueName: \"kubernetes.io/projected/01193382-e383-4185-bbd6-af046b634069-kube-api-access-t9vkk\") on node \"crc\" DevicePath \"\"" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.426875 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.426904 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-32e9-account-create-update-f8gtl" event={"ID":"4f11741e-da8f-4bb7-abf3-038efc7ee2a6","Type":"ContainerDied","Data":"f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511"} Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.427378 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49db7da22465bdcf133788923bb98a5fd1bab9d0df69af5881913715dac5511" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.429634 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-j2kwl" event={"ID":"01193382-e383-4185-bbd6-af046b634069","Type":"ContainerDied","Data":"f22546f09145bdf9f35ed211a8538e7e2cec792835c80bea7c5159dd79071c5d"} Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.429669 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f22546f09145bdf9f35ed211a8538e7e2cec792835c80bea7c5159dd79071c5d" Nov 25 23:15:52 crc kubenswrapper[4761]: I1125 23:15:52.429749 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-j2kwl" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.649336 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-l2dbz"] Nov 25 23:15:53 crc kubenswrapper[4761]: E1125 23:15:53.649786 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f11741e-da8f-4bb7-abf3-038efc7ee2a6" containerName="mariadb-account-create-update" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.649845 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f11741e-da8f-4bb7-abf3-038efc7ee2a6" containerName="mariadb-account-create-update" Nov 25 23:15:53 crc kubenswrapper[4761]: E1125 23:15:53.649875 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01193382-e383-4185-bbd6-af046b634069" containerName="mariadb-database-create" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.649889 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="01193382-e383-4185-bbd6-af046b634069" containerName="mariadb-database-create" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.650118 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f11741e-da8f-4bb7-abf3-038efc7ee2a6" containerName="mariadb-account-create-update" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.650148 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="01193382-e383-4185-bbd6-af046b634069" containerName="mariadb-database-create" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.650912 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.655691 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.655815 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-76m6w" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.670891 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-l2dbz"] Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.690520 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.690568 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.690595 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tgzn\" (UniqueName: \"kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.791577 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.791620 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.792268 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tgzn\" (UniqueName: \"kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.798966 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.804292 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.823373 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tgzn\" (UniqueName: \"kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn\") pod \"glance-db-sync-l2dbz\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:53 crc kubenswrapper[4761]: I1125 23:15:53.996390 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:54 crc kubenswrapper[4761]: I1125 23:15:54.453832 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-l2dbz"] Nov 25 23:15:55 crc kubenswrapper[4761]: I1125 23:15:55.462030 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-l2dbz" event={"ID":"06655c0a-237e-4a10-b264-55c63afb1198","Type":"ContainerStarted","Data":"28aa62ad597020bbf52bcb8be8d3e3c0ad243c54f981e1d57babddda42fc93f0"} Nov 25 23:15:55 crc kubenswrapper[4761]: I1125 23:15:55.462617 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-l2dbz" event={"ID":"06655c0a-237e-4a10-b264-55c63afb1198","Type":"ContainerStarted","Data":"60958e8243037109cabf2e8dc6f64f70d2dc8fd8e61b38df726e1f27b89431a4"} Nov 25 23:15:55 crc kubenswrapper[4761]: I1125 23:15:55.494347 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-l2dbz" podStartSLOduration=2.494322546 podStartE2EDuration="2.494322546s" podCreationTimestamp="2025-11-25 23:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:15:55.487899444 +0000 UTC m=+1151.221245329" watchObservedRunningTime="2025-11-25 23:15:55.494322546 +0000 UTC m=+1151.227668411" Nov 25 23:15:58 crc kubenswrapper[4761]: I1125 23:15:58.527507 4761 generic.go:334] "Generic (PLEG): container finished" podID="06655c0a-237e-4a10-b264-55c63afb1198" containerID="28aa62ad597020bbf52bcb8be8d3e3c0ad243c54f981e1d57babddda42fc93f0" exitCode=0 Nov 25 23:15:58 crc kubenswrapper[4761]: I1125 23:15:58.527659 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-l2dbz" event={"ID":"06655c0a-237e-4a10-b264-55c63afb1198","Type":"ContainerDied","Data":"28aa62ad597020bbf52bcb8be8d3e3c0ad243c54f981e1d57babddda42fc93f0"} Nov 25 23:15:59 crc kubenswrapper[4761]: I1125 23:15:59.949836 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:15:59 crc kubenswrapper[4761]: I1125 23:15:59.993648 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data\") pod \"06655c0a-237e-4a10-b264-55c63afb1198\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " Nov 25 23:15:59 crc kubenswrapper[4761]: I1125 23:15:59.993759 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data\") pod \"06655c0a-237e-4a10-b264-55c63afb1198\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " Nov 25 23:15:59 crc kubenswrapper[4761]: I1125 23:15:59.993807 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tgzn\" (UniqueName: \"kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn\") pod \"06655c0a-237e-4a10-b264-55c63afb1198\" (UID: \"06655c0a-237e-4a10-b264-55c63afb1198\") " Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.001465 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "06655c0a-237e-4a10-b264-55c63afb1198" (UID: "06655c0a-237e-4a10-b264-55c63afb1198"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.001956 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn" (OuterVolumeSpecName: "kube-api-access-8tgzn") pod "06655c0a-237e-4a10-b264-55c63afb1198" (UID: "06655c0a-237e-4a10-b264-55c63afb1198"). InnerVolumeSpecName "kube-api-access-8tgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.048796 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data" (OuterVolumeSpecName: "config-data") pod "06655c0a-237e-4a10-b264-55c63afb1198" (UID: "06655c0a-237e-4a10-b264-55c63afb1198"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.095237 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.095272 4761 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06655c0a-237e-4a10-b264-55c63afb1198-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.095286 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tgzn\" (UniqueName: \"kubernetes.io/projected/06655c0a-237e-4a10-b264-55c63afb1198-kube-api-access-8tgzn\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.551080 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-l2dbz" event={"ID":"06655c0a-237e-4a10-b264-55c63afb1198","Type":"ContainerDied","Data":"60958e8243037109cabf2e8dc6f64f70d2dc8fd8e61b38df726e1f27b89431a4"} Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.551136 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60958e8243037109cabf2e8dc6f64f70d2dc8fd8e61b38df726e1f27b89431a4" Nov 25 23:16:00 crc kubenswrapper[4761]: I1125 23:16:00.551236 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-l2dbz" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.756528 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:16:01 crc kubenswrapper[4761]: E1125 23:16:01.757067 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06655c0a-237e-4a10-b264-55c63afb1198" containerName="glance-db-sync" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.757079 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="06655c0a-237e-4a10-b264-55c63afb1198" containerName="glance-db-sync" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.757197 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="06655c0a-237e-4a10-b264-55c63afb1198" containerName="glance-db-sync" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.758071 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.760482 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.760653 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.761248 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-76m6w" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.782091 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.821494 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923051 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923102 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nrsq\" (UniqueName: \"kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923146 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923174 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923198 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923218 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923238 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923410 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923464 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923554 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923673 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923757 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923864 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.923934 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.924288 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:01 crc kubenswrapper[4761]: I1125 23:16:01.942853 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.009018 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.010778 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.013896 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025167 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025223 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025250 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025294 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025331 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025341 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025382 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025356 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nrsq\" (UniqueName: \"kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025467 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025328 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025522 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025499 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025607 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025671 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025691 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025724 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025753 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025769 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025788 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025857 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.025897 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.026091 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.026110 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.030238 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.033156 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.048345 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.051837 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.060399 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nrsq\" (UniqueName: \"kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq\") pod \"glance-default-external-api-0\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.116645 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127592 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127649 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127681 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127790 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127823 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127842 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127860 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127877 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127891 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47xgf\" (UniqueName: \"kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127920 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127940 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127957 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.127986 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.128005 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.229180 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.229278 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.229309 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.229498 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.229312 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230196 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230222 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47xgf\" (UniqueName: \"kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230240 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230266 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230290 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230317 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230346 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230371 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230403 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230425 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230446 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230485 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.230746 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.231001 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.231054 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.231548 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.231915 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.231960 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.232015 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.232236 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.238158 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.243081 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.253758 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47xgf\" (UniqueName: \"kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.258175 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.279654 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.329552 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.535869 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:16:02 crc kubenswrapper[4761]: W1125 23:16:02.536202 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod498370d5_ca44_4d35_89b2_9870dff0374a.slice/crio-038fd72f5bea03661fa2050cda00e684f280c6fdf0b1bb616c4d6d1bf70c63b7 WatchSource:0}: Error finding container 038fd72f5bea03661fa2050cda00e684f280c6fdf0b1bb616c4d6d1bf70c63b7: Status 404 returned error can't find the container with id 038fd72f5bea03661fa2050cda00e684f280c6fdf0b1bb616c4d6d1bf70c63b7 Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.565503 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerStarted","Data":"038fd72f5bea03661fa2050cda00e684f280c6fdf0b1bb616c4d6d1bf70c63b7"} Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.631807 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:02 crc kubenswrapper[4761]: I1125 23:16:02.756843 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:02 crc kubenswrapper[4761]: W1125 23:16:02.769018 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7389966_a269_459a_bbe1_72d852cc8fff.slice/crio-516b27b867519f7d7b34774c7bf00235297fd41d9b418bcf166f377a620e17f6 WatchSource:0}: Error finding container 516b27b867519f7d7b34774c7bf00235297fd41d9b418bcf166f377a620e17f6: Status 404 returned error can't find the container with id 516b27b867519f7d7b34774c7bf00235297fd41d9b418bcf166f377a620e17f6 Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.577083 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerStarted","Data":"5910c6a426abd2c113d99b56a40667340d1976fbe384d403a6ec96fedbdfcbb3"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.577813 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerStarted","Data":"302494bd6d6e85ec65176c3fbc7331edaa2c254340853f2248669e146ca33d9b"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.577833 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerStarted","Data":"3131aee74aa040be6a00b9572c38e93263950b5f35b9f14361d2e59fde5da6a0"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.582747 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerStarted","Data":"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.582792 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerStarted","Data":"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.582802 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerStarted","Data":"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.582810 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerStarted","Data":"516b27b867519f7d7b34774c7bf00235297fd41d9b418bcf166f377a620e17f6"} Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.582993 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-log" containerID="cri-o://658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" gracePeriod=30 Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.583025 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-httpd" containerID="cri-o://c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" gracePeriod=30 Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.583017 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-api" containerID="cri-o://54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" gracePeriod=30 Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.634189 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.63416243 podStartE2EDuration="2.63416243s" podCreationTimestamp="2025-11-25 23:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:16:03.615338861 +0000 UTC m=+1159.348684736" watchObservedRunningTime="2025-11-25 23:16:03.63416243 +0000 UTC m=+1159.367508295" Nov 25 23:16:03 crc kubenswrapper[4761]: I1125 23:16:03.657643 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.657616794 podStartE2EDuration="3.657616794s" podCreationTimestamp="2025-11-25 23:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:16:03.651382397 +0000 UTC m=+1159.384728242" watchObservedRunningTime="2025-11-25 23:16:03.657616794 +0000 UTC m=+1159.390962659" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.056872 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.163852 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47xgf\" (UniqueName: \"kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.163936 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.163988 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164031 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164067 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164095 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164138 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164162 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164197 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164171 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164226 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys" (OuterVolumeSpecName: "sys") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164233 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164187 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164246 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164199 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run" (OuterVolumeSpecName: "run") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164269 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev" (OuterVolumeSpecName: "dev") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164289 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164387 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164397 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164423 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164461 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick\") pod \"b7389966-a269-459a-bbe1-72d852cc8fff\" (UID: \"b7389966-a269-459a-bbe1-72d852cc8fff\") " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.164663 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs" (OuterVolumeSpecName: "logs") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165018 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165029 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165038 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165047 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165055 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165063 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165070 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165079 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7389966-a269-459a-bbe1-72d852cc8fff-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.165104 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.169136 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf" (OuterVolumeSpecName: "kube-api-access-47xgf") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "kube-api-access-47xgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.169355 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.170789 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts" (OuterVolumeSpecName: "scripts") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.171844 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.247465 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data" (OuterVolumeSpecName: "config-data") pod "b7389966-a269-459a-bbe1-72d852cc8fff" (UID: "b7389966-a269-459a-bbe1-72d852cc8fff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266619 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266719 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266754 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266775 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7389966-a269-459a-bbe1-72d852cc8fff-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266797 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47xgf\" (UniqueName: \"kubernetes.io/projected/b7389966-a269-459a-bbe1-72d852cc8fff-kube-api-access-47xgf\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.266814 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7389966-a269-459a-bbe1-72d852cc8fff-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.279078 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.279294 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.368316 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.368389 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.598662 4761 generic.go:334] "Generic (PLEG): container finished" podID="b7389966-a269-459a-bbe1-72d852cc8fff" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" exitCode=143 Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.598877 4761 generic.go:334] "Generic (PLEG): container finished" podID="b7389966-a269-459a-bbe1-72d852cc8fff" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" exitCode=143 Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.598969 4761 generic.go:334] "Generic (PLEG): container finished" podID="b7389966-a269-459a-bbe1-72d852cc8fff" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" exitCode=143 Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.598767 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerDied","Data":"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e"} Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.598766 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.599096 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerDied","Data":"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af"} Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.599266 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerDied","Data":"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e"} Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.599124 4761 scope.go:117] "RemoveContainer" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.599334 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7389966-a269-459a-bbe1-72d852cc8fff","Type":"ContainerDied","Data":"516b27b867519f7d7b34774c7bf00235297fd41d9b418bcf166f377a620e17f6"} Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.638850 4761 scope.go:117] "RemoveContainer" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.663335 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.670786 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.671503 4761 scope.go:117] "RemoveContainer" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.710442 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.710793 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-api" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.710818 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-api" Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.710848 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-log" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.710857 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-log" Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.710876 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-httpd" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.710884 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-httpd" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.711042 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-httpd" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.711060 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-api" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.711078 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" containerName="glance-log" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.712251 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.712264 4761 scope.go:117] "RemoveContainer" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.712866 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": container with ID starting with 54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e not found: ID does not exist" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.712905 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e"} err="failed to get container status \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": rpc error: code = NotFound desc = could not find container \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": container with ID starting with 54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.712932 4761 scope.go:117] "RemoveContainer" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.714074 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": container with ID starting with c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af not found: ID does not exist" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.714107 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af"} err="failed to get container status \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": rpc error: code = NotFound desc = could not find container \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": container with ID starting with c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.714150 4761 scope.go:117] "RemoveContainer" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.714320 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.718547 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:04 crc kubenswrapper[4761]: E1125 23:16:04.721649 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": container with ID starting with 658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e not found: ID does not exist" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.721845 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e"} err="failed to get container status \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": rpc error: code = NotFound desc = could not find container \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": container with ID starting with 658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.722005 4761 scope.go:117] "RemoveContainer" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.724868 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e"} err="failed to get container status \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": rpc error: code = NotFound desc = could not find container \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": container with ID starting with 54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.724913 4761 scope.go:117] "RemoveContainer" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725189 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af"} err="failed to get container status \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": rpc error: code = NotFound desc = could not find container \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": container with ID starting with c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725202 4761 scope.go:117] "RemoveContainer" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725433 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e"} err="failed to get container status \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": rpc error: code = NotFound desc = could not find container \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": container with ID starting with 658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725445 4761 scope.go:117] "RemoveContainer" containerID="54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725612 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e"} err="failed to get container status \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": rpc error: code = NotFound desc = could not find container \"54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e\": container with ID starting with 54feed45411d11e6a720a8a703464a360b54c628481b3ed85375cf21c9ccb24e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725623 4761 scope.go:117] "RemoveContainer" containerID="c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725800 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af"} err="failed to get container status \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": rpc error: code = NotFound desc = could not find container \"c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af\": container with ID starting with c3e281ec5738cc26c4ef34987bed1dec971eb5c1967c94daf501ef58859e85af not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725811 4761 scope.go:117] "RemoveContainer" containerID="658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.725974 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e"} err="failed to get container status \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": rpc error: code = NotFound desc = could not find container \"658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e\": container with ID starting with 658a75cfe9bca0ac78fa2e7c4163fcc52481a6bbae3ec241a9f3e830cfbc1a2e not found: ID does not exist" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.774885 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775230 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775273 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775299 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775321 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775434 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775504 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775549 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775605 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775731 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775787 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775816 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jcvv\" (UniqueName: \"kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775896 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.775937 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877024 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877093 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877125 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877145 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877162 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877160 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877180 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877215 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877266 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877291 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877198 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877382 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877312 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877487 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877488 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877522 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877547 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877573 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jcvv\" (UniqueName: \"kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877597 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877615 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877733 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877774 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877809 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.877870 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.878030 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.884616 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.888335 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.894910 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jcvv\" (UniqueName: \"kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.903728 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:04 crc kubenswrapper[4761]: I1125 23:16:04.911140 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:05 crc kubenswrapper[4761]: I1125 23:16:05.029999 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7389966-a269-459a-bbe1-72d852cc8fff" path="/var/lib/kubelet/pods/b7389966-a269-459a-bbe1-72d852cc8fff/volumes" Nov 25 23:16:05 crc kubenswrapper[4761]: I1125 23:16:05.075319 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:05 crc kubenswrapper[4761]: I1125 23:16:05.333264 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:16:05 crc kubenswrapper[4761]: W1125 23:16:05.341035 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf66b46b7_0b13_4ae9_a945_ee8d8f424492.slice/crio-c32da9efae5ec1ab37ef0ee926acdf1a78d103c49f689b160b3fe8622bd57901 WatchSource:0}: Error finding container c32da9efae5ec1ab37ef0ee926acdf1a78d103c49f689b160b3fe8622bd57901: Status 404 returned error can't find the container with id c32da9efae5ec1ab37ef0ee926acdf1a78d103c49f689b160b3fe8622bd57901 Nov 25 23:16:05 crc kubenswrapper[4761]: I1125 23:16:05.616285 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerStarted","Data":"9c7761d642885dd8874a392f404719008ac20fa5e8e4a6cd42e432c165875c45"} Nov 25 23:16:05 crc kubenswrapper[4761]: I1125 23:16:05.616996 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerStarted","Data":"c32da9efae5ec1ab37ef0ee926acdf1a78d103c49f689b160b3fe8622bd57901"} Nov 25 23:16:06 crc kubenswrapper[4761]: I1125 23:16:06.638299 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerStarted","Data":"0cf9efa34066c0d5e1299152b15e28d04ca1fd68c8c07f9d0e3b179e5da98363"} Nov 25 23:16:06 crc kubenswrapper[4761]: I1125 23:16:06.638785 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerStarted","Data":"d1debbdaa27b4f5b6e8586b21ec3bac136b5dde89aa9a793489a29937c1cf78b"} Nov 25 23:16:06 crc kubenswrapper[4761]: I1125 23:16:06.693799 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.693768629 podStartE2EDuration="2.693768629s" podCreationTimestamp="2025-11-25 23:16:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:16:06.682669834 +0000 UTC m=+1162.416015729" watchObservedRunningTime="2025-11-25 23:16:06.693768629 +0000 UTC m=+1162.427114494" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.117321 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.118167 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.118190 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.162247 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.163962 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.191249 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.699263 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.699614 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.699626 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.711811 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.714973 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:12 crc kubenswrapper[4761]: I1125 23:16:12.718786 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.076011 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.076312 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.076327 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.116577 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.124166 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.143046 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.729935 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.730288 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.730428 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.750970 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.751592 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:15 crc kubenswrapper[4761]: I1125 23:16:15.755823 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:16:17 crc kubenswrapper[4761]: I1125 23:16:17.970902 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:16:17 crc kubenswrapper[4761]: I1125 23:16:17.974223 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:17 crc kubenswrapper[4761]: I1125 23:16:17.979062 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:16:17 crc kubenswrapper[4761]: I1125 23:16:17.980791 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:17 crc kubenswrapper[4761]: I1125 23:16:17.999910 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013087 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013153 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013193 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013240 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013275 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013304 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013330 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013365 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013393 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013433 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.013476 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014023 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014097 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014166 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014209 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014244 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014280 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014351 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014380 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014438 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014487 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014525 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014571 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.014905 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.015832 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp8hr\" (UniqueName: \"kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.015904 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.015943 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.015974 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsclt\" (UniqueName: \"kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.028971 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117458 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117554 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117594 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117627 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117656 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117749 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117764 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117823 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117862 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117791 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.117888 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118157 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118238 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118423 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118560 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118595 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118632 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118672 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118724 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118750 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118801 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118819 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118810 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118869 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.118943 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119064 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119217 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119299 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119388 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119467 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119554 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119240 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119595 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119668 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119807 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119828 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.119883 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120012 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120035 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120058 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp8hr\" (UniqueName: \"kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120101 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120138 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120167 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsclt\" (UniqueName: \"kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120245 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120254 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120284 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120321 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120457 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120467 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.120494 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.127821 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.131119 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.131524 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.141354 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.144472 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.148095 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp8hr\" (UniqueName: \"kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.150147 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsclt\" (UniqueName: \"kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.160259 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.170235 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.174529 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.305190 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.320590 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.697467 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:16:18 crc kubenswrapper[4761]: W1125 23:16:18.713050 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab76fb_74be_4d5b_9c82_e932b210d1c5.slice/crio-f44a95a25d29d9797625fe0ea3a807c70316ca6fbcad716461637bb3f8d464db WatchSource:0}: Error finding container f44a95a25d29d9797625fe0ea3a807c70316ca6fbcad716461637bb3f8d464db: Status 404 returned error can't find the container with id f44a95a25d29d9797625fe0ea3a807c70316ca6fbcad716461637bb3f8d464db Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.754718 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerStarted","Data":"f44a95a25d29d9797625fe0ea3a807c70316ca6fbcad716461637bb3f8d464db"} Nov 25 23:16:18 crc kubenswrapper[4761]: W1125 23:16:18.977826 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05e189be_0a23_4c59_8f4d_9c32e8a0cb40.slice/crio-8b9a78a3732d2ce19bf50e9523fb0eb9b08936b13cac292264fc59be0b9daadd WatchSource:0}: Error finding container 8b9a78a3732d2ce19bf50e9523fb0eb9b08936b13cac292264fc59be0b9daadd: Status 404 returned error can't find the container with id 8b9a78a3732d2ce19bf50e9523fb0eb9b08936b13cac292264fc59be0b9daadd Nov 25 23:16:18 crc kubenswrapper[4761]: I1125 23:16:18.980769 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.770468 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerStarted","Data":"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.771448 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerStarted","Data":"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.771480 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerStarted","Data":"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.771498 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerStarted","Data":"8b9a78a3732d2ce19bf50e9523fb0eb9b08936b13cac292264fc59be0b9daadd"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.774023 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerStarted","Data":"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.774069 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerStarted","Data":"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.774089 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerStarted","Data":"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72"} Nov 25 23:16:19 crc kubenswrapper[4761]: I1125 23:16:19.822346 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.8223083410000003 podStartE2EDuration="3.822308341s" podCreationTimestamp="2025-11-25 23:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:16:19.819844766 +0000 UTC m=+1175.553190651" watchObservedRunningTime="2025-11-25 23:16:19.822308341 +0000 UTC m=+1175.555654216" Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.128109 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.128879 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.128960 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.129920 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.130022 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09" gracePeriod=600 Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.794245 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09" exitCode=0 Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.794360 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09"} Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.794670 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115"} Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.794713 4761 scope.go:117] "RemoveContainer" containerID="9211d7b5e5c39e7e35b3a9f683b867d6921e5bb146c5bb13aa5a92412988a4a8" Nov 25 23:16:21 crc kubenswrapper[4761]: I1125 23:16:21.812834 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=5.81281295 podStartE2EDuration="5.81281295s" podCreationTimestamp="2025-11-25 23:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:16:19.863295908 +0000 UTC m=+1175.596641813" watchObservedRunningTime="2025-11-25 23:16:21.81281295 +0000 UTC m=+1177.546158785" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.306081 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.306894 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.306918 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.321746 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.322048 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.322061 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.359979 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.361205 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.364515 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.366114 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.372159 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.374403 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.876580 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.876640 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.876956 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.876985 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.877006 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.877029 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.892562 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.893760 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.895285 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.895347 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.896001 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:16:28 crc kubenswrapper[4761]: I1125 23:16:28.900328 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:18:21 crc kubenswrapper[4761]: I1125 23:18:21.128576 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:18:21 crc kubenswrapper[4761]: I1125 23:18:21.129318 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:18:51 crc kubenswrapper[4761]: I1125 23:18:51.128225 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:18:51 crc kubenswrapper[4761]: I1125 23:18:51.128954 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.127813 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.128491 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.128554 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.129353 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.129465 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115" gracePeriod=600 Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.580071 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115" exitCode=0 Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.580172 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115"} Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.580538 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251"} Nov 25 23:19:21 crc kubenswrapper[4761]: I1125 23:19:21.580572 4761 scope.go:117] "RemoveContainer" containerID="42156081f6cac07111dc677e9f948c0d6bf9c1abc5ca9e72b884de8ff08dfa09" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.334121 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.336394 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.339558 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.340579 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.402520 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.494748 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.495027 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.596661 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.596915 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.597051 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.629066 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.704932 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:44 crc kubenswrapper[4761]: I1125 23:19:44.986663 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 23:19:45 crc kubenswrapper[4761]: I1125 23:19:45.824290 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fd190a1c-2058-4ac4-8217-68b61fced6d4","Type":"ContainerStarted","Data":"211a7ee7108f8bdec585c0114cdb24ea4a5dff2d052adb9199d437171b2af8f9"} Nov 25 23:19:45 crc kubenswrapper[4761]: I1125 23:19:45.824627 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fd190a1c-2058-4ac4-8217-68b61fced6d4","Type":"ContainerStarted","Data":"c3a25421710785ac74f540cad8c50f31a77bdd00f2a0462a4d6919e24f655f9a"} Nov 25 23:19:45 crc kubenswrapper[4761]: I1125 23:19:45.843692 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.84367265 podStartE2EDuration="1.84367265s" podCreationTimestamp="2025-11-25 23:19:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:19:45.839967661 +0000 UTC m=+1381.573313496" watchObservedRunningTime="2025-11-25 23:19:45.84367265 +0000 UTC m=+1381.577018505" Nov 25 23:19:46 crc kubenswrapper[4761]: I1125 23:19:46.833227 4761 generic.go:334] "Generic (PLEG): container finished" podID="fd190a1c-2058-4ac4-8217-68b61fced6d4" containerID="211a7ee7108f8bdec585c0114cdb24ea4a5dff2d052adb9199d437171b2af8f9" exitCode=0 Nov 25 23:19:46 crc kubenswrapper[4761]: I1125 23:19:46.833540 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fd190a1c-2058-4ac4-8217-68b61fced6d4","Type":"ContainerDied","Data":"211a7ee7108f8bdec585c0114cdb24ea4a5dff2d052adb9199d437171b2af8f9"} Nov 25 23:19:46 crc kubenswrapper[4761]: I1125 23:19:46.908140 4761 scope.go:117] "RemoveContainer" containerID="9633597078bb507c73766fef261263383998d4db894a50ee8c63b8e499c59573" Nov 25 23:19:46 crc kubenswrapper[4761]: I1125 23:19:46.938388 4761 scope.go:117] "RemoveContainer" containerID="ce50da77323e94f7c89db224b8eb05c4265e5f6d6a1bb23b2c689baad23a602b" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.306124 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.307874 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir\") pod \"fd190a1c-2058-4ac4-8217-68b61fced6d4\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.308050 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access\") pod \"fd190a1c-2058-4ac4-8217-68b61fced6d4\" (UID: \"fd190a1c-2058-4ac4-8217-68b61fced6d4\") " Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.308084 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fd190a1c-2058-4ac4-8217-68b61fced6d4" (UID: "fd190a1c-2058-4ac4-8217-68b61fced6d4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.309948 4761 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fd190a1c-2058-4ac4-8217-68b61fced6d4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.315425 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fd190a1c-2058-4ac4-8217-68b61fced6d4" (UID: "fd190a1c-2058-4ac4-8217-68b61fced6d4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.411533 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd190a1c-2058-4ac4-8217-68b61fced6d4-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.855937 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fd190a1c-2058-4ac4-8217-68b61fced6d4","Type":"ContainerDied","Data":"c3a25421710785ac74f540cad8c50f31a77bdd00f2a0462a4d6919e24f655f9a"} Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.855990 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3a25421710785ac74f540cad8c50f31a77bdd00f2a0462a4d6919e24f655f9a" Nov 25 23:19:48 crc kubenswrapper[4761]: I1125 23:19:48.856023 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.133488 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 23:19:50 crc kubenswrapper[4761]: E1125 23:19:50.134010 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd190a1c-2058-4ac4-8217-68b61fced6d4" containerName="pruner" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.134032 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd190a1c-2058-4ac4-8217-68b61fced6d4" containerName="pruner" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.134303 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd190a1c-2058-4ac4-8217-68b61fced6d4" containerName="pruner" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.135213 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.142259 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.142586 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.142829 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.142973 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.150544 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.199352 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.244115 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.244191 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.244295 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.244393 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.244719 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.271933 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access\") pod \"installer-9-crc\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:50 crc kubenswrapper[4761]: I1125 23:19:50.514252 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:19:51 crc kubenswrapper[4761]: I1125 23:19:51.068322 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 23:19:51 crc kubenswrapper[4761]: I1125 23:19:51.889564 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6568a4b2-1689-4cfa-84a7-1d952014beac","Type":"ContainerStarted","Data":"f8685399dc35014a9edc28f913f2cf87e2eb4717605bbf4319abf1496a0f8d99"} Nov 25 23:19:51 crc kubenswrapper[4761]: I1125 23:19:51.889637 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6568a4b2-1689-4cfa-84a7-1d952014beac","Type":"ContainerStarted","Data":"5c92733a9bcc4218d9c28746053071464a9caced8ff9cad189fb9ba095bf7c9c"} Nov 25 23:19:51 crc kubenswrapper[4761]: I1125 23:19:51.916901 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.916879855 podStartE2EDuration="1.916879855s" podCreationTimestamp="2025-11-25 23:19:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:19:51.909286592 +0000 UTC m=+1387.642632437" watchObservedRunningTime="2025-11-25 23:19:51.916879855 +0000 UTC m=+1387.650225690" Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.187116 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.188796 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-log" containerID="cri-o://d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.188947 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-api" containerID="cri-o://d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.188991 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-httpd" containerID="cri-o://c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.203602 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.203975 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-log" containerID="cri-o://1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.204161 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-api" containerID="cri-o://57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.204242 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-httpd" containerID="cri-o://3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" gracePeriod=30 Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.875106 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:20:20 crc kubenswrapper[4761]: I1125 23:20:20.888356 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036070 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036112 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036134 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036158 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036198 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036214 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036229 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036252 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036271 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036289 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036304 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036333 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036349 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036378 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036392 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036406 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036421 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036435 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036450 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036465 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036518 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp8hr\" (UniqueName: \"kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036533 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036568 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsclt\" (UniqueName: \"kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036590 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036611 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036631 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run\") pod \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\" (UID: \"05e189be-0a23-4c59-8f4d-9c32e8a0cb40\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036648 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036672 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme\") pod \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\" (UID: \"69ab76fb-74be-4d5b-9c82-e932b210d1c5\") " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036784 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev" (OuterVolumeSpecName: "dev") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036897 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.036997 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037016 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037212 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037294 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037340 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs" (OuterVolumeSpecName: "logs") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037378 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run" (OuterVolumeSpecName: "run") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037400 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys" (OuterVolumeSpecName: "sys") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037423 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037423 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037475 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037449 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run" (OuterVolumeSpecName: "run") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037509 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys" (OuterVolumeSpecName: "sys") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037534 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037784 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037812 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037840 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.037882 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev" (OuterVolumeSpecName: "dev") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.038121 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs" (OuterVolumeSpecName: "logs") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.042280 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.042871 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts" (OuterVolumeSpecName: "scripts") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.043758 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.045401 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.045437 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr" (OuterVolumeSpecName: "kube-api-access-zp8hr") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "kube-api-access-zp8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.046796 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts" (OuterVolumeSpecName: "scripts") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.048563 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt" (OuterVolumeSpecName: "kube-api-access-lsclt") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "kube-api-access-lsclt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.049083 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.124921 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data" (OuterVolumeSpecName: "config-data") pod "69ab76fb-74be-4d5b-9c82-e932b210d1c5" (UID: "69ab76fb-74be-4d5b-9c82-e932b210d1c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.135162 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data" (OuterVolumeSpecName: "config-data") pod "05e189be-0a23-4c59-8f4d-9c32e8a0cb40" (UID: "05e189be-0a23-4c59-8f4d-9c32e8a0cb40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139315 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139362 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139408 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139428 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139452 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139472 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139490 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139507 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139524 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139542 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139558 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139578 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139596 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139627 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139653 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139673 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp8hr\" (UniqueName: \"kubernetes.io/projected/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-kube-api-access-zp8hr\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139730 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsclt\" (UniqueName: \"kubernetes.io/projected/69ab76fb-74be-4d5b-9c82-e932b210d1c5-kube-api-access-lsclt\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139751 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69ab76fb-74be-4d5b-9c82-e932b210d1c5-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139769 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139787 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139804 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139823 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/69ab76fb-74be-4d5b-9c82-e932b210d1c5-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139851 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139869 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139887 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69ab76fb-74be-4d5b-9c82-e932b210d1c5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.139904 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/05e189be-0a23-4c59-8f4d-9c32e8a0cb40-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.162292 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166848 4761 generic.go:334] "Generic (PLEG): container finished" podID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" exitCode=0 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166899 4761 generic.go:334] "Generic (PLEG): container finished" podID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" exitCode=0 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166913 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166942 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerDied","Data":"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166987 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerDied","Data":"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.167007 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerDied","Data":"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.166920 4761 generic.go:334] "Generic (PLEG): container finished" podID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" exitCode=143 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.167034 4761 scope.go:117] "RemoveContainer" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.167341 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"69ab76fb-74be-4d5b-9c82-e932b210d1c5","Type":"ContainerDied","Data":"f44a95a25d29d9797625fe0ea3a807c70316ca6fbcad716461637bb3f8d464db"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.168625 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170403 4761 generic.go:334] "Generic (PLEG): container finished" podID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" exitCode=0 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170436 4761 generic.go:334] "Generic (PLEG): container finished" podID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" exitCode=0 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170447 4761 generic.go:334] "Generic (PLEG): container finished" podID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" exitCode=143 Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170467 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerDied","Data":"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170488 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerDied","Data":"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170504 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerDied","Data":"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170543 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170518 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"05e189be-0a23-4c59-8f4d-9c32e8a0cb40","Type":"ContainerDied","Data":"8b9a78a3732d2ce19bf50e9523fb0eb9b08936b13cac292264fc59be0b9daadd"} Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.170814 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.176999 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.212217 4761 scope.go:117] "RemoveContainer" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.222330 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.233816 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.240905 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.240982 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.241200 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.241293 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.241366 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.249451 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.250162 4761 scope.go:117] "RemoveContainer" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.273315 4761 scope.go:117] "RemoveContainer" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.273868 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": container with ID starting with d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6 not found: ID does not exist" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.273922 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6"} err="failed to get container status \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": rpc error: code = NotFound desc = could not find container \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": container with ID starting with d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.273956 4761 scope.go:117] "RemoveContainer" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.274351 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": container with ID starting with c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd not found: ID does not exist" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.274474 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd"} err="failed to get container status \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": rpc error: code = NotFound desc = could not find container \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": container with ID starting with c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.274583 4761 scope.go:117] "RemoveContainer" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.276145 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": container with ID starting with d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72 not found: ID does not exist" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.276192 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72"} err="failed to get container status \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": rpc error: code = NotFound desc = could not find container \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": container with ID starting with d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.276223 4761 scope.go:117] "RemoveContainer" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.276739 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6"} err="failed to get container status \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": rpc error: code = NotFound desc = could not find container \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": container with ID starting with d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.276759 4761 scope.go:117] "RemoveContainer" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.276997 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd"} err="failed to get container status \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": rpc error: code = NotFound desc = could not find container \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": container with ID starting with c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277018 4761 scope.go:117] "RemoveContainer" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277280 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72"} err="failed to get container status \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": rpc error: code = NotFound desc = could not find container \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": container with ID starting with d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277299 4761 scope.go:117] "RemoveContainer" containerID="d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277560 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6"} err="failed to get container status \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": rpc error: code = NotFound desc = could not find container \"d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6\": container with ID starting with d128c3cb5b46e34aff8185ccee1330a12d1be1ccbaa79396863a2f8575e840d6 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277586 4761 scope.go:117] "RemoveContainer" containerID="c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277893 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd"} err="failed to get container status \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": rpc error: code = NotFound desc = could not find container \"c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd\": container with ID starting with c4d03113f5c0f346aa3827b080555249891d8575d75a9a15ae22c126cd90d8cd not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.277921 4761 scope.go:117] "RemoveContainer" containerID="d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.278130 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72"} err="failed to get container status \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": rpc error: code = NotFound desc = could not find container \"d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72\": container with ID starting with d89311298d79f340202e7edab2ca554877aaf9cb2c56cbb5f9df7665b0f18d72 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.278155 4761 scope.go:117] "RemoveContainer" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.302876 4761 scope.go:117] "RemoveContainer" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.372355 4761 scope.go:117] "RemoveContainer" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.391631 4761 scope.go:117] "RemoveContainer" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.393140 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": container with ID starting with 57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc not found: ID does not exist" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.393181 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc"} err="failed to get container status \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": rpc error: code = NotFound desc = could not find container \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": container with ID starting with 57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.393212 4761 scope.go:117] "RemoveContainer" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.393633 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": container with ID starting with 3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e not found: ID does not exist" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.393758 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e"} err="failed to get container status \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": rpc error: code = NotFound desc = could not find container \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": container with ID starting with 3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.393872 4761 scope.go:117] "RemoveContainer" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" Nov 25 23:20:21 crc kubenswrapper[4761]: E1125 23:20:21.394419 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": container with ID starting with 1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487 not found: ID does not exist" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.394451 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487"} err="failed to get container status \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": rpc error: code = NotFound desc = could not find container \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": container with ID starting with 1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.394487 4761 scope.go:117] "RemoveContainer" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.394923 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc"} err="failed to get container status \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": rpc error: code = NotFound desc = could not find container \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": container with ID starting with 57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.395024 4761 scope.go:117] "RemoveContainer" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.395477 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e"} err="failed to get container status \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": rpc error: code = NotFound desc = could not find container \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": container with ID starting with 3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.395585 4761 scope.go:117] "RemoveContainer" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.396021 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487"} err="failed to get container status \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": rpc error: code = NotFound desc = could not find container \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": container with ID starting with 1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487 not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.396119 4761 scope.go:117] "RemoveContainer" containerID="57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.396533 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc"} err="failed to get container status \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": rpc error: code = NotFound desc = could not find container \"57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc\": container with ID starting with 57b61918dc2698284f10d5ea78039bcd5762d8aee540f44ebfaade0499f73fbc not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.396561 4761 scope.go:117] "RemoveContainer" containerID="3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.397018 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e"} err="failed to get container status \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": rpc error: code = NotFound desc = could not find container \"3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e\": container with ID starting with 3182e7505a3d0ffddd9c7a98729f6c0bd3fe579d6967a2559dcfdbc48db6d56e not found: ID does not exist" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.397075 4761 scope.go:117] "RemoveContainer" containerID="1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487" Nov 25 23:20:21 crc kubenswrapper[4761]: I1125 23:20:21.397373 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487"} err="failed to get container status \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": rpc error: code = NotFound desc = could not find container \"1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487\": container with ID starting with 1d6b73ef551ba096eacdb2b872c2a4caa93dc2d3adc17638da40da031dad4487 not found: ID does not exist" Nov 25 23:20:23 crc kubenswrapper[4761]: I1125 23:20:23.036780 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" path="/var/lib/kubelet/pods/05e189be-0a23-4c59-8f4d-9c32e8a0cb40/volumes" Nov 25 23:20:23 crc kubenswrapper[4761]: I1125 23:20:23.039080 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" path="/var/lib/kubelet/pods/69ab76fb-74be-4d5b-9c82-e932b210d1c5/volumes" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.983282 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984021 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984042 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984094 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984107 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984129 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984143 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984165 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984177 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984212 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984224 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: E1125 23:20:26.984249 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984261 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984484 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984508 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984523 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-httpd" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984550 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-api" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984567 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ab76fb-74be-4d5b-9c82-e932b210d1c5" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.984582 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e189be-0a23-4c59-8f4d-9c32e8a0cb40" containerName="glance-log" Nov 25 23:20:26 crc kubenswrapper[4761]: I1125 23:20:26.986375 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.008970 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.038020 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhtg5\" (UniqueName: \"kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.039676 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.039788 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.141331 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.141378 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.141453 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhtg5\" (UniqueName: \"kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.144650 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.144908 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.173416 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhtg5\" (UniqueName: \"kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5\") pod \"redhat-marketplace-dtxht\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.324070 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:27 crc kubenswrapper[4761]: I1125 23:20:27.560776 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:20:28 crc kubenswrapper[4761]: I1125 23:20:28.232447 4761 generic.go:334] "Generic (PLEG): container finished" podID="346df741-cf07-4002-a8b6-5069c015aee3" containerID="2de507af97056a10556e8e772f957e4454dd1625c3a3dd19f2c4398f0d25ca58" exitCode=0 Nov 25 23:20:28 crc kubenswrapper[4761]: I1125 23:20:28.232535 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerDied","Data":"2de507af97056a10556e8e772f957e4454dd1625c3a3dd19f2c4398f0d25ca58"} Nov 25 23:20:28 crc kubenswrapper[4761]: I1125 23:20:28.234072 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerStarted","Data":"d9ffda992202ddfb7687df31ddd1d88513d106f1270ec7e81cee31de28e8e083"} Nov 25 23:20:28 crc kubenswrapper[4761]: I1125 23:20:28.235475 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.244406 4761 generic.go:334] "Generic (PLEG): container finished" podID="346df741-cf07-4002-a8b6-5069c015aee3" containerID="500c5ae5a21baefe2028e90c608b4187f9e3bed40d1b7bac534a7eda3bc0b93f" exitCode=0 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.244449 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerDied","Data":"500c5ae5a21baefe2028e90c608b4187f9e3bed40d1b7bac534a7eda3bc0b93f"} Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.269112 4761 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270120 4761 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270281 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270424 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657" gracePeriod=15 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270524 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028" gracePeriod=15 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270617 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2" gracePeriod=15 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270543 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9" gracePeriod=15 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.270844 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318" gracePeriod=15 Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276472 4761 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.276852 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276874 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.276899 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276908 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.276938 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276947 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.276958 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276966 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.276983 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.276992 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.277006 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277014 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.277029 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277038 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277220 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277245 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277261 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277274 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277290 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.277304 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.342613 4761 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381008 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381228 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381333 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381406 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381471 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381551 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381624 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.381731 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.482868 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483288 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483329 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483347 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483397 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483043 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483446 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483458 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483533 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483574 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483617 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483685 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483643 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483813 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.483907 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.484072 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.665313 4761 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.665644 4761 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.665919 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.666178 4761 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.667028 4761 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.667867 4761 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:29 crc kubenswrapper[4761]: I1125 23:20:29.667922 4761 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.668274 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="200ms" Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.671343 4761 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-dtxht.187b63459c2984c6 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-dtxht,UID:346df741-cf07-4002-a8b6-5069c015aee3,APIVersion:v1,ResourceVersion:42447,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 423ms (423ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 23:20:29.669541062 +0000 UTC m=+1425.402886907,LastTimestamp:2025-11-25 23:20:29.669541062 +0000 UTC m=+1425.402886907,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 23:20:29 crc kubenswrapper[4761]: W1125 23:20:29.713971 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-8fbfd5479e6f5da980b97b2d5e0764ede456db3cc4c4def409189f34014017bd WatchSource:0}: Error finding container 8fbfd5479e6f5da980b97b2d5e0764ede456db3cc4c4def409189f34014017bd: Status 404 returned error can't find the container with id 8fbfd5479e6f5da980b97b2d5e0764ede456db3cc4c4def409189f34014017bd Nov 25 23:20:29 crc kubenswrapper[4761]: E1125 23:20:29.869683 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="400ms" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.257341 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.261099 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.261969 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028" exitCode=0 Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.262017 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2" exitCode=0 Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.262036 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318" exitCode=0 Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.262055 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9" exitCode=2 Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.262099 4761 scope.go:117] "RemoveContainer" containerID="0f51db0aa6f561f7d0b2e19c9e58738a4eae9cd41d46867cf5cf9f1dc21c99d3" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.264584 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09"} Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.264667 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"8fbfd5479e6f5da980b97b2d5e0764ede456db3cc4c4def409189f34014017bd"} Nov 25 23:20:30 crc kubenswrapper[4761]: E1125 23:20:30.265668 4761 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.266144 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.268575 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerStarted","Data":"960bb5300b34104f250886cdbc3862a2dd991ea88b658211150c1e5819545b24"} Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.269541 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.270204 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: E1125 23:20:30.270303 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="800ms" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.270532 4761 generic.go:334] "Generic (PLEG): container finished" podID="6568a4b2-1689-4cfa-84a7-1d952014beac" containerID="f8685399dc35014a9edc28f913f2cf87e2eb4717605bbf4319abf1496a0f8d99" exitCode=0 Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.270590 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6568a4b2-1689-4cfa-84a7-1d952014beac","Type":"ContainerDied","Data":"f8685399dc35014a9edc28f913f2cf87e2eb4717605bbf4319abf1496a0f8d99"} Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.271257 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.271739 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.272255 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.472100 4761 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 23:20:30 crc kubenswrapper[4761]: I1125 23:20:30.472267 4761 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: E1125 23:20:31.071392 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="1.6s" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.284662 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.771188 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.772342 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.772690 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.777671 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.778651 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.779044 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.779387 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.780011 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848162 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848273 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848594 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock\") pod \"6568a4b2-1689-4cfa-84a7-1d952014beac\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848670 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848712 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir\") pod \"6568a4b2-1689-4cfa-84a7-1d952014beac\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848817 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access\") pod \"6568a4b2-1689-4cfa-84a7-1d952014beac\" (UID: \"6568a4b2-1689-4cfa-84a7-1d952014beac\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848843 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848909 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6568a4b2-1689-4cfa-84a7-1d952014beac" (UID: "6568a4b2-1689-4cfa-84a7-1d952014beac"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.848962 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.849124 4761 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.849134 4761 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.849142 4761 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.849159 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.849447 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock" (OuterVolumeSpecName: "var-lock") pod "6568a4b2-1689-4cfa-84a7-1d952014beac" (UID: "6568a4b2-1689-4cfa-84a7-1d952014beac"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.857356 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6568a4b2-1689-4cfa-84a7-1d952014beac" (UID: "6568a4b2-1689-4cfa-84a7-1d952014beac"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.951207 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6568a4b2-1689-4cfa-84a7-1d952014beac-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.951255 4761 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:31 crc kubenswrapper[4761]: I1125 23:20:31.951268 4761 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6568a4b2-1689-4cfa-84a7-1d952014beac-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.035107 4761 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/mysql-db-openstack-galera-1: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/mysql-db-openstack-galera-1\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="glance-kuttl-tests/openstack-galera-1" volumeName="mysql-db" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.302237 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.304285 4761 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657" exitCode=0 Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.304430 4761 scope.go:117] "RemoveContainer" containerID="c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.304446 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.307969 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6568a4b2-1689-4cfa-84a7-1d952014beac","Type":"ContainerDied","Data":"5c92733a9bcc4218d9c28746053071464a9caced8ff9cad189fb9ba095bf7c9c"} Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.308022 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c92733a9bcc4218d9c28746053071464a9caced8ff9cad189fb9ba095bf7c9c" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.308173 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.341482 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.342774 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.343302 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.348378 4761 scope.go:117] "RemoveContainer" containerID="8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.361565 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.362264 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.362808 4761 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.383694 4761 scope.go:117] "RemoveContainer" containerID="a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.407165 4761 scope.go:117] "RemoveContainer" containerID="af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.440608 4761 scope.go:117] "RemoveContainer" containerID="2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.482559 4761 scope.go:117] "RemoveContainer" containerID="a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.519517 4761 scope.go:117] "RemoveContainer" containerID="c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.520491 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\": container with ID starting with c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028 not found: ID does not exist" containerID="c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.520561 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028"} err="failed to get container status \"c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\": rpc error: code = NotFound desc = could not find container \"c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028\": container with ID starting with c149ea3b9b32890fe6b9fd6c223d397ca518a204e5ca1e5d6e35d9072491c028 not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.520610 4761 scope.go:117] "RemoveContainer" containerID="8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.521115 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\": container with ID starting with 8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2 not found: ID does not exist" containerID="8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.521153 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2"} err="failed to get container status \"8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\": rpc error: code = NotFound desc = could not find container \"8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2\": container with ID starting with 8d4b11ca485f7bd2c60cbd02330a13fdc2135320e0e4305a5357bcf56c3ccec2 not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.521182 4761 scope.go:117] "RemoveContainer" containerID="a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.522045 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\": container with ID starting with a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318 not found: ID does not exist" containerID="a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.522140 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318"} err="failed to get container status \"a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\": rpc error: code = NotFound desc = could not find container \"a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318\": container with ID starting with a2f2116a98d51b9ea56b877ecc4cc35acc6f2bcb4bf42b750ec9bba51c855318 not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.522160 4761 scope.go:117] "RemoveContainer" containerID="af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.522776 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\": container with ID starting with af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9 not found: ID does not exist" containerID="af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.522803 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9"} err="failed to get container status \"af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\": rpc error: code = NotFound desc = could not find container \"af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9\": container with ID starting with af5a8648f76ce0bdb89dcd71107bd01b22556f13c673699875ae9b815d18a4a9 not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.522821 4761 scope.go:117] "RemoveContainer" containerID="2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.523175 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\": container with ID starting with 2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657 not found: ID does not exist" containerID="2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.523243 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657"} err="failed to get container status \"2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\": rpc error: code = NotFound desc = could not find container \"2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657\": container with ID starting with 2ef4d7a862a96aa74b2cacf0367b517ecd2c93cf2eb956a9bac93f3a14367657 not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.523287 4761 scope.go:117] "RemoveContainer" containerID="a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.523668 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\": container with ID starting with a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc not found: ID does not exist" containerID="a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc" Nov 25 23:20:32 crc kubenswrapper[4761]: I1125 23:20:32.523727 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc"} err="failed to get container status \"a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\": rpc error: code = NotFound desc = could not find container \"a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc\": container with ID starting with a57d04e4182b82e496bef5be192f6951b9e39c2a1189f3a6f9a1aa0784e1ccdc not found: ID does not exist" Nov 25 23:20:32 crc kubenswrapper[4761]: E1125 23:20:32.672873 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="3.2s" Nov 25 23:20:33 crc kubenswrapper[4761]: I1125 23:20:33.025366 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 23:20:35 crc kubenswrapper[4761]: I1125 23:20:35.021337 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:35 crc kubenswrapper[4761]: I1125 23:20:35.022426 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:35 crc kubenswrapper[4761]: E1125 23:20:35.873425 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="6.4s" Nov 25 23:20:36 crc kubenswrapper[4761]: E1125 23:20:36.013800 4761 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/glance-glance-default-internal-api-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/glance-glance-default-internal-api-0\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="glance-kuttl-tests/glance-default-internal-api-0" volumeName="glance" Nov 25 23:20:36 crc kubenswrapper[4761]: E1125 23:20:36.014474 4761 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/glance-cache-glance-default-internal-api-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/glance-cache-glance-default-internal-api-0\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="glance-kuttl-tests/glance-default-internal-api-0" volumeName="glance-cache" Nov 25 23:20:37 crc kubenswrapper[4761]: E1125 23:20:37.034568 4761 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/swift-swift-storage-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/swift-swift-storage-0\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="glance-kuttl-tests/swift-storage-0" volumeName="swift" Nov 25 23:20:37 crc kubenswrapper[4761]: I1125 23:20:37.324889 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:37 crc kubenswrapper[4761]: I1125 23:20:37.325391 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:37 crc kubenswrapper[4761]: I1125 23:20:37.397840 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:37 crc kubenswrapper[4761]: I1125 23:20:37.398651 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:37 crc kubenswrapper[4761]: I1125 23:20:37.399388 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:38 crc kubenswrapper[4761]: E1125 23:20:38.305414 4761 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-dtxht.187b63459c2984c6 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-dtxht,UID:346df741-cf07-4002-a8b6-5069c015aee3,APIVersion:v1,ResourceVersion:42447,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 423ms (423ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 23:20:29.669541062 +0000 UTC m=+1425.402886907,LastTimestamp:2025-11-25 23:20:29.669541062 +0000 UTC m=+1425.402886907,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 23:20:38 crc kubenswrapper[4761]: I1125 23:20:38.449655 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:20:38 crc kubenswrapper[4761]: I1125 23:20:38.450452 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:38 crc kubenswrapper[4761]: I1125 23:20:38.451106 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.385277 4761 generic.go:334] "Generic (PLEG): container finished" podID="91b5b48e-a642-45e2-a84a-d449123e57bb" containerID="bf334415f6cd4cfd9d90cdd581fa39798b88c952ac303a041dae5a20705e849e" exitCode=1 Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.385395 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerDied","Data":"bf334415f6cd4cfd9d90cdd581fa39798b88c952ac303a041dae5a20705e849e"} Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.386062 4761 scope.go:117] "RemoveContainer" containerID="bf334415f6cd4cfd9d90cdd581fa39798b88c952ac303a041dae5a20705e849e" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.386758 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.387343 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.388354 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.389002 4761 generic.go:334] "Generic (PLEG): container finished" podID="47a4708b-be53-4783-b9c1-9536d90b9e4c" containerID="cde14783c38bcdf6cfb028f2cebbd11e85d7de90cd9950c41d54501602a21e86" exitCode=1 Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.389458 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerDied","Data":"cde14783c38bcdf6cfb028f2cebbd11e85d7de90cd9950c41d54501602a21e86"} Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.391052 4761 scope.go:117] "RemoveContainer" containerID="cde14783c38bcdf6cfb028f2cebbd11e85d7de90cd9950c41d54501602a21e86" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.391247 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.392312 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.393003 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:39 crc kubenswrapper[4761]: I1125 23:20:39.393474 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.401835 4761 generic.go:334] "Generic (PLEG): container finished" podID="2e34c701-6bbc-4000-a90a-aeca2333b69b" containerID="9567de82eec7b39af5f0974fe4a4370e520ded7e082979e868d5438cfa39451a" exitCode=1 Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.401897 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerDied","Data":"9567de82eec7b39af5f0974fe4a4370e520ded7e082979e868d5438cfa39451a"} Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.404051 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.404086 4761 scope.go:117] "RemoveContainer" containerID="9567de82eec7b39af5f0974fe4a4370e520ded7e082979e868d5438cfa39451a" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.404471 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.404913 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.405326 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.405827 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.406895 4761 generic.go:334] "Generic (PLEG): container finished" podID="91b5b48e-a642-45e2-a84a-d449123e57bb" containerID="906547d42b27288c36b83091debb1e5648cac010f5d1d98c2f3ac19207225689" exitCode=1 Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.407030 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerDied","Data":"906547d42b27288c36b83091debb1e5648cac010f5d1d98c2f3ac19207225689"} Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.407098 4761 scope.go:117] "RemoveContainer" containerID="bf334415f6cd4cfd9d90cdd581fa39798b88c952ac303a041dae5a20705e849e" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.407860 4761 scope.go:117] "RemoveContainer" containerID="906547d42b27288c36b83091debb1e5648cac010f5d1d98c2f3ac19207225689" Nov 25 23:20:40 crc kubenswrapper[4761]: E1125 23:20:40.408248 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-cskvg_openstack-operators(91b5b48e-a642-45e2-a84a-d449123e57bb)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.408536 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.409275 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.410386 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.411498 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.412668 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.414627 4761 generic.go:334] "Generic (PLEG): container finished" podID="47a4708b-be53-4783-b9c1-9536d90b9e4c" containerID="775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35" exitCode=1 Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.414754 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerDied","Data":"775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35"} Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.416017 4761 scope.go:117] "RemoveContainer" containerID="775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35" Nov 25 23:20:40 crc kubenswrapper[4761]: E1125 23:20:40.416509 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.417161 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.419433 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.420015 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.420623 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.421078 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:40 crc kubenswrapper[4761]: I1125 23:20:40.476268 4761 scope.go:117] "RemoveContainer" containerID="cde14783c38bcdf6cfb028f2cebbd11e85d7de90cd9950c41d54501602a21e86" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.427932 4761 generic.go:334] "Generic (PLEG): container finished" podID="bca3f822-b2ce-401e-b219-52d29e761c49" containerID="61d3817e1eb40c038638d44c35a68380ac5d3f980e9129cbc735b3a11dbc55a3" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.428064 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerDied","Data":"61d3817e1eb40c038638d44c35a68380ac5d3f980e9129cbc735b3a11dbc55a3"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.429121 4761 scope.go:117] "RemoveContainer" containerID="61d3817e1eb40c038638d44c35a68380ac5d3f980e9129cbc735b3a11dbc55a3" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.429482 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.429787 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.430241 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.430793 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.431285 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.431575 4761 generic.go:334] "Generic (PLEG): container finished" podID="8aba4458-8023-4b9f-8504-6fab49f5250c" containerID="3d10673a484790e525cda35f8eb9fc95ab6f6dbc67bbaa621dbbe35ec976a66b" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.431617 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerDied","Data":"3d10673a484790e525cda35f8eb9fc95ab6f6dbc67bbaa621dbbe35ec976a66b"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.431888 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.432512 4761 scope.go:117] "RemoveContainer" containerID="3d10673a484790e525cda35f8eb9fc95ab6f6dbc67bbaa621dbbe35ec976a66b" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.432654 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.433191 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.433567 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.433829 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.434116 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.434379 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.434780 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.435005 4761 generic.go:334] "Generic (PLEG): container finished" podID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" containerID="80ddb3e8634b4d68c9f701a1c14b8be0415779a56d956cee8b895f3962cdda58" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.435098 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerDied","Data":"80ddb3e8634b4d68c9f701a1c14b8be0415779a56d956cee8b895f3962cdda58"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.435625 4761 scope.go:117] "RemoveContainer" containerID="80ddb3e8634b4d68c9f701a1c14b8be0415779a56d956cee8b895f3962cdda58" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.436358 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.436971 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.437351 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.438105 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.438580 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.439174 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.439565 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.439984 4761 generic.go:334] "Generic (PLEG): container finished" podID="2e34c701-6bbc-4000-a90a-aeca2333b69b" containerID="828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.440064 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerDied","Data":"828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.440129 4761 scope.go:117] "RemoveContainer" containerID="9567de82eec7b39af5f0974fe4a4370e520ded7e082979e868d5438cfa39451a" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.440758 4761 scope.go:117] "RemoveContainer" containerID="828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d" Nov 25 23:20:41 crc kubenswrapper[4761]: E1125 23:20:41.441161 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-5c5dd6c96-z8jr7_openstack-operators(2e34c701-6bbc-4000-a90a-aeca2333b69b)\"" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.442028 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.442688 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.443247 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.444209 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.445510 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.445797 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.446320 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.447141 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.447617 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.452232 4761 generic.go:334] "Generic (PLEG): container finished" podID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" containerID="2ba5de066a343f8a191d00b472cce29283efeea1f98421fb4d1e5a4a3e7e0fb9" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.452354 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerDied","Data":"2ba5de066a343f8a191d00b472cce29283efeea1f98421fb4d1e5a4a3e7e0fb9"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.453306 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.453360 4761 scope.go:117] "RemoveContainer" containerID="2ba5de066a343f8a191d00b472cce29283efeea1f98421fb4d1e5a4a3e7e0fb9" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.453869 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.454390 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.454812 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.455004 4761 generic.go:334] "Generic (PLEG): container finished" podID="0717f7d1-d5b3-408e-878e-10c62251448a" containerID="31d225e37138e67eda4e68aee35ffa55d0cc8952747907f352a150fd953f45ff" exitCode=1 Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.455049 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerDied","Data":"31d225e37138e67eda4e68aee35ffa55d0cc8952747907f352a150fd953f45ff"} Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.455249 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.455524 4761 scope.go:117] "RemoveContainer" containerID="31d225e37138e67eda4e68aee35ffa55d0cc8952747907f352a150fd953f45ff" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.455996 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.456429 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.456967 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.457450 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.458579 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.459520 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.460097 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.460469 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.461000 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.461650 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.462156 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.462595 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.463093 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:41 crc kubenswrapper[4761]: I1125 23:20:41.463476 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.010576 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.012693 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.013409 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.014137 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.014670 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.015236 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.015753 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.016339 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.016889 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.017354 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.017881 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.161385 4761 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.161851 4761 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.162583 4761 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.163409 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.275223 4761 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="7s" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.476481 4761 generic.go:334] "Generic (PLEG): container finished" podID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" containerID="c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531" exitCode=1 Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.476605 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerDied","Data":"c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.478113 4761 scope.go:117] "RemoveContainer" containerID="2ba5de066a343f8a191d00b472cce29283efeea1f98421fb4d1e5a4a3e7e0fb9" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.479128 4761 scope.go:117] "RemoveContainer" containerID="c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.479568 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-66b769fbc8-xf6tp_openstack-operators(3daf70a9-4dab-4ace-82c0-9c7e48512cf6)\"" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.479690 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.480225 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.480803 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.481393 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.481543 4761 generic.go:334] "Generic (PLEG): container finished" podID="0717f7d1-d5b3-408e-878e-10c62251448a" containerID="3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89" exitCode=1 Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.481616 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerDied","Data":"3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.481953 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.482297 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.482614 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.482726 4761 scope.go:117] "RemoveContainer" containerID="3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.483010 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.483104 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-6bbbcc9f68-dcp42_openstack-operators(0717f7d1-d5b3-408e-878e-10c62251448a)\"" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.483360 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.483870 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.484349 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.484782 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.485501 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.485995 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.486350 4761 generic.go:334] "Generic (PLEG): container finished" podID="bca3f822-b2ce-401e-b219-52d29e761c49" containerID="8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527" exitCode=1 Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.486383 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerDied","Data":"8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.486923 4761 scope.go:117] "RemoveContainer" containerID="8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.487239 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-6d87477c79-vr8fq_metallb-system(bca3f822-b2ce-401e-b219-52d29e761c49)\"" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.487441 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.487859 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.488250 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.488647 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.489126 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.489578 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.489792 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.490067 4761 generic.go:334] "Generic (PLEG): container finished" podID="8aba4458-8023-4b9f-8504-6fab49f5250c" containerID="cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7" exitCode=1 Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.490146 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerDied","Data":"cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.490978 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.491094 4761 scope.go:117] "RemoveContainer" containerID="cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.491771 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.492232 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.492776 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.493219 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.493670 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.494109 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.494506 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.494968 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.495333 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.495730 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.496047 4761 generic.go:334] "Generic (PLEG): container finished" podID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" containerID="7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7" exitCode=1 Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.496174 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.496388 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerDied","Data":"7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.496490 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.496885 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.497042 4761 scope.go:117] "RemoveContainer" containerID="7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7" Nov 25 23:20:42 crc kubenswrapper[4761]: E1125 23:20:42.497588 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-7f54679d79-6ldpn_openstack-operators(bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b)\"" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.497642 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.498060 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.498450 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.498881 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.499054 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f58f331945b65f16aea8b5fe1016c65df89263ae3a00c6baf26dbb922745249c"} Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.499976 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.500451 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.501287 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.502097 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.502579 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.503079 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.503608 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.505526 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.506000 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.506494 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.506993 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.507406 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.507912 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.548315 4761 scope.go:117] "RemoveContainer" containerID="31d225e37138e67eda4e68aee35ffa55d0cc8952747907f352a150fd953f45ff" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.660734 4761 scope.go:117] "RemoveContainer" containerID="61d3817e1eb40c038638d44c35a68380ac5d3f980e9129cbc735b3a11dbc55a3" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.726134 4761 scope.go:117] "RemoveContainer" containerID="3d10673a484790e525cda35f8eb9fc95ab6f6dbc67bbaa621dbbe35ec976a66b" Nov 25 23:20:42 crc kubenswrapper[4761]: I1125 23:20:42.773791 4761 scope.go:117] "RemoveContainer" containerID="80ddb3e8634b4d68c9f701a1c14b8be0415779a56d956cee8b895f3962cdda58" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.522333 4761 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d8d5430f723ecc3a2808b5833c2994bd2473d5a951b5b241833962416dac0092" exitCode=0 Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.522486 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d8d5430f723ecc3a2808b5833c2994bd2473d5a951b5b241833962416dac0092"} Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.523054 4761 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.523092 4761 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:43 crc kubenswrapper[4761]: E1125 23:20:43.523844 4761 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.524275 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.525251 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.526295 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.526350 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.526374 4761 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4" exitCode=1 Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.526403 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4"} Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.527828 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.530758 4761 scope.go:117] "RemoveContainer" containerID="67739b37eb982aaa2e25afe184c413ee574a4a39505a0d28fd46c38dcd04c6b4" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.531024 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.531481 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.531934 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.532643 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.533186 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.533899 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.534477 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.534959 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.535405 4761 scope.go:117] "RemoveContainer" containerID="3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.535499 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: E1125 23:20:43.535791 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-6bbbcc9f68-dcp42_openstack-operators(0717f7d1-d5b3-408e-878e-10c62251448a)\"" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.536081 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.536538 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.537022 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.537678 4761 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.538389 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.539076 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.539759 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.540412 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.541250 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.541628 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.542120 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.542855 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.543360 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.543979 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.544460 4761 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.544951 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.545452 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.545927 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.546530 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.674495 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.675452 4761 scope.go:117] "RemoveContainer" containerID="828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.675584 4761 status_manager.go:851] "Failed to get status for pod" podUID="346df741-cf07-4002-a8b6-5069c015aee3" pod="openshift-marketplace/redhat-marketplace-dtxht" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dtxht\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: E1125 23:20:43.675874 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-5c5dd6c96-z8jr7_openstack-operators(2e34c701-6bbc-4000-a90a-aeca2333b69b)\"" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.676385 4761 status_manager.go:851] "Failed to get status for pod" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c5dd6c96-z8jr7\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.677597 4761 status_manager.go:851] "Failed to get status for pod" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-66b769fbc8-xf6tp\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.678302 4761 status_manager.go:851] "Failed to get status for pod" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-6c6d6c68d5-v5cfc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.679119 4761 status_manager.go:851] "Failed to get status for pod" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-6bbbcc9f68-dcp42\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.679612 4761 status_manager.go:851] "Failed to get status for pod" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-cskvg\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.680301 4761 status_manager.go:851] "Failed to get status for pod" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-7f54679d79-6ldpn\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.681056 4761 status_manager.go:851] "Failed to get status for pod" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-8559b887bb-frnqv\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.681537 4761 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.682113 4761 status_manager.go:851] "Failed to get status for pod" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:43 crc kubenswrapper[4761]: I1125 23:20:43.682604 4761 status_manager.go:851] "Failed to get status for pod" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d87477c79-vr8fq\": dial tcp 38.102.83.73:6443: connect: connection refused" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.216095 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.216649 4761 scope.go:117] "RemoveContainer" containerID="cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7" Nov 25 23:20:44 crc kubenswrapper[4761]: E1125 23:20:44.216877 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.548338 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.548639 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cd80afac9624becd2189217e6f0ba3fe7f6d5d81e164251c3034f99e72615815"} Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.551618 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"77d4316c82c4c70c292e5c2f775658095d9cef211e1668c8457d4e64275b5505"} Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.551657 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"01f117afbf2cea25195b6f207fee734b7e10fc072076982f1afa3fedfa502721"} Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.551667 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3a5807b19c568f030478a4da3bd9eda3bdf16acadd568dca2d1a01cf16fdd833"} Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.820741 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.821268 4761 scope.go:117] "RemoveContainer" containerID="7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7" Nov 25 23:20:44 crc kubenswrapper[4761]: E1125 23:20:44.821480 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-7f54679d79-6ldpn_openstack-operators(bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b)\"" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.841263 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.841313 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:20:44 crc kubenswrapper[4761]: I1125 23:20:44.841874 4761 scope.go:117] "RemoveContainer" containerID="775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35" Nov 25 23:20:44 crc kubenswrapper[4761]: E1125 23:20:44.842079 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:20:45 crc kubenswrapper[4761]: I1125 23:20:45.561825 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"654943ff423acd3be5a6259d6217aa86e5753c1e1372ddd7f2e8156ecc3fd930"} Nov 25 23:20:45 crc kubenswrapper[4761]: I1125 23:20:45.562859 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e21d19ea2962584e21e417b88593824c35a9bc99d2bc9fd350fb4b95e0c4d1de"} Nov 25 23:20:45 crc kubenswrapper[4761]: I1125 23:20:45.562958 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:45 crc kubenswrapper[4761]: I1125 23:20:45.562819 4761 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:45 crc kubenswrapper[4761]: I1125 23:20:45.563204 4761 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:47 crc kubenswrapper[4761]: I1125 23:20:47.017449 4761 scope.go:117] "RemoveContainer" containerID="bd97a4391c11a978c36a4883f1ce7ade43cf0bcd91e22909f420dfb79e198787" Nov 25 23:20:47 crc kubenswrapper[4761]: I1125 23:20:47.164518 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:47 crc kubenswrapper[4761]: I1125 23:20:47.164663 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:47 crc kubenswrapper[4761]: I1125 23:20:47.171322 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:48 crc kubenswrapper[4761]: I1125 23:20:48.037663 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 23:20:48 crc kubenswrapper[4761]: I1125 23:20:48.047265 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 23:20:48 crc kubenswrapper[4761]: I1125 23:20:48.439460 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 23:20:49 crc kubenswrapper[4761]: I1125 23:20:49.604839 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:20:49 crc kubenswrapper[4761]: I1125 23:20:49.605876 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:20:49 crc kubenswrapper[4761]: I1125 23:20:49.606686 4761 scope.go:117] "RemoveContainer" containerID="c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531" Nov 25 23:20:49 crc kubenswrapper[4761]: E1125 23:20:49.607093 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-66b769fbc8-xf6tp_openstack-operators(3daf70a9-4dab-4ace-82c0-9c7e48512cf6)\"" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.004300 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.005313 4761 scope.go:117] "RemoveContainer" containerID="8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527" Nov 25 23:20:50 crc kubenswrapper[4761]: E1125 23:20:50.005791 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-6d87477c79-vr8fq_metallb-system(bca3f822-b2ce-401e-b219-52d29e761c49)\"" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.575491 4761 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.607454 4761 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.608034 4761 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.617142 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:20:50 crc kubenswrapper[4761]: I1125 23:20:50.742898 4761 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5ecc47b4-679b-40cd-958f-30b80324ff76" Nov 25 23:20:51 crc kubenswrapper[4761]: I1125 23:20:51.615026 4761 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:51 crc kubenswrapper[4761]: I1125 23:20:51.615354 4761 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="feda2a26-8d6f-4e04-b8b3-f606cd5408c5" Nov 25 23:20:51 crc kubenswrapper[4761]: I1125 23:20:51.618322 4761 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5ecc47b4-679b-40cd-958f-30b80324ff76" Nov 25 23:20:52 crc kubenswrapper[4761]: I1125 23:20:52.490193 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:20:52 crc kubenswrapper[4761]: I1125 23:20:52.491212 4761 scope.go:117] "RemoveContainer" containerID="3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89" Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.642630 4761 generic.go:334] "Generic (PLEG): container finished" podID="0717f7d1-d5b3-408e-878e-10c62251448a" containerID="545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6" exitCode=1 Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.642680 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerDied","Data":"545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6"} Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.643022 4761 scope.go:117] "RemoveContainer" containerID="3a20a45b0bf0aa7c0fc984f1f02cbccabb2c247e0a2c01c7eacea2fbfb431e89" Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.643918 4761 scope.go:117] "RemoveContainer" containerID="545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6" Nov 25 23:20:53 crc kubenswrapper[4761]: E1125 23:20:53.644330 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-6bbbcc9f68-dcp42_openstack-operators(0717f7d1-d5b3-408e-878e-10c62251448a)\"" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.674949 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:20:53 crc kubenswrapper[4761]: I1125 23:20:53.675982 4761 scope.go:117] "RemoveContainer" containerID="828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.215400 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.216381 4761 scope.go:117] "RemoveContainer" containerID="cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.655531 4761 generic.go:334] "Generic (PLEG): container finished" podID="2e34c701-6bbc-4000-a90a-aeca2333b69b" containerID="edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96" exitCode=1 Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.655613 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerDied","Data":"edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96"} Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.656014 4761 scope.go:117] "RemoveContainer" containerID="828566abe2de3d187b24f69db2a828d871d0238a8675a61ec81d3fa69e1c811d" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.657096 4761 scope.go:117] "RemoveContainer" containerID="edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96" Nov 25 23:20:54 crc kubenswrapper[4761]: E1125 23:20:54.658154 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c5dd6c96-z8jr7_openstack-operators(2e34c701-6bbc-4000-a90a-aeca2333b69b)\"" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.666540 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerStarted","Data":"89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce"} Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.667174 4761 scope.go:117] "RemoveContainer" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" Nov 25 23:20:54 crc kubenswrapper[4761]: E1125 23:20:54.667450 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.820155 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:20:54 crc kubenswrapper[4761]: I1125 23:20:54.820680 4761 scope.go:117] "RemoveContainer" containerID="7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7" Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.683677 4761 generic.go:334] "Generic (PLEG): container finished" podID="8aba4458-8023-4b9f-8504-6fab49f5250c" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" exitCode=1 Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.683760 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerDied","Data":"89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce"} Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.683951 4761 scope.go:117] "RemoveContainer" containerID="cd5f8bdb8dbf8b5386b4e88182f9199adbec6094d514a17b98024322fabd98c7" Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.684912 4761 scope.go:117] "RemoveContainer" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" Nov 25 23:20:55 crc kubenswrapper[4761]: E1125 23:20:55.685543 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.687611 4761 generic.go:334] "Generic (PLEG): container finished" podID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" containerID="c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22" exitCode=1 Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.687673 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerDied","Data":"c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22"} Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.688582 4761 scope.go:117] "RemoveContainer" containerID="c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22" Nov 25 23:20:55 crc kubenswrapper[4761]: E1125 23:20:55.688990 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-7f54679d79-6ldpn_openstack-operators(bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b)\"" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" Nov 25 23:20:55 crc kubenswrapper[4761]: I1125 23:20:55.776653 4761 scope.go:117] "RemoveContainer" containerID="7414a06357c9679085820ceb55ac175930fee87440c9fdbe262be698fcd449c7" Nov 25 23:20:56 crc kubenswrapper[4761]: I1125 23:20:56.012316 4761 scope.go:117] "RemoveContainer" containerID="906547d42b27288c36b83091debb1e5648cac010f5d1d98c2f3ac19207225689" Nov 25 23:20:56 crc kubenswrapper[4761]: I1125 23:20:56.708288 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerStarted","Data":"82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8"} Nov 25 23:20:57 crc kubenswrapper[4761]: I1125 23:20:57.725594 4761 generic.go:334] "Generic (PLEG): container finished" podID="91b5b48e-a642-45e2-a84a-d449123e57bb" containerID="82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8" exitCode=1 Nov 25 23:20:57 crc kubenswrapper[4761]: I1125 23:20:57.725671 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerDied","Data":"82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8"} Nov 25 23:20:57 crc kubenswrapper[4761]: I1125 23:20:57.725773 4761 scope.go:117] "RemoveContainer" containerID="906547d42b27288c36b83091debb1e5648cac010f5d1d98c2f3ac19207225689" Nov 25 23:20:57 crc kubenswrapper[4761]: I1125 23:20:57.726548 4761 scope.go:117] "RemoveContainer" containerID="82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8" Nov 25 23:20:57 crc kubenswrapper[4761]: E1125 23:20:57.727025 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-cskvg_openstack-operators(91b5b48e-a642-45e2-a84a-d449123e57bb)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.012765 4761 scope.go:117] "RemoveContainer" containerID="775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35" Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.446188 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.737386 4761 generic.go:334] "Generic (PLEG): container finished" podID="47a4708b-be53-4783-b9c1-9536d90b9e4c" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" exitCode=1 Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.737429 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerDied","Data":"30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f"} Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.737459 4761 scope.go:117] "RemoveContainer" containerID="775dd4dedc5a160a7f02c3214bda7e3eb94ff95b28061f8812ad908c668e1a35" Nov 25 23:20:58 crc kubenswrapper[4761]: I1125 23:20:58.738035 4761 scope.go:117] "RemoveContainer" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" Nov 25 23:20:58 crc kubenswrapper[4761]: E1125 23:20:58.738359 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:20:59 crc kubenswrapper[4761]: I1125 23:20:59.850895 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 23:21:00 crc kubenswrapper[4761]: I1125 23:21:00.051791 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 23:21:00 crc kubenswrapper[4761]: I1125 23:21:00.732766 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 23:21:00 crc kubenswrapper[4761]: I1125 23:21:00.909401 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 23:21:01 crc kubenswrapper[4761]: I1125 23:21:01.060823 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 23:21:01 crc kubenswrapper[4761]: I1125 23:21:01.637373 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.011371 4761 scope.go:117] "RemoveContainer" containerID="8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.011528 4761 scope.go:117] "RemoveContainer" containerID="c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.114209 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.173598 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.270182 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.279602 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.333759 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.436286 4761 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.444195 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dtxht" podStartSLOduration=35.009722036 podStartE2EDuration="36.444142433s" podCreationTimestamp="2025-11-25 23:20:26 +0000 UTC" firstStartedPulling="2025-11-25 23:20:28.235107235 +0000 UTC m=+1423.968453110" lastFinishedPulling="2025-11-25 23:20:29.669527652 +0000 UTC m=+1425.402873507" observedRunningTime="2025-11-25 23:20:50.657814384 +0000 UTC m=+1446.391160269" watchObservedRunningTime="2025-11-25 23:21:02.444142433 +0000 UTC m=+1458.177488308" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.451907 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.451998 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.457371 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.474725 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.474673921 podStartE2EDuration="12.474673921s" podCreationTimestamp="2025-11-25 23:20:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:21:02.471805264 +0000 UTC m=+1458.205151119" watchObservedRunningTime="2025-11-25 23:21:02.474673921 +0000 UTC m=+1458.208019766" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.490203 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.491181 4761 scope.go:117] "RemoveContainer" containerID="545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6" Nov 25 23:21:02 crc kubenswrapper[4761]: E1125 23:21:02.491657 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-6bbbcc9f68-dcp42_openstack-operators(0717f7d1-d5b3-408e-878e-10c62251448a)\"" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.500273 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.538005 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.580044 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.599835 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2cpjb" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.642485 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.679212 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.709157 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.736613 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.783280 4761 generic.go:334] "Generic (PLEG): container finished" podID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" containerID="b0246e180b5baf24dd4b04c8a155233673df8f15aca70ad693f75b747959e68e" exitCode=1 Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.783532 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerDied","Data":"b0246e180b5baf24dd4b04c8a155233673df8f15aca70ad693f75b747959e68e"} Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.783586 4761 scope.go:117] "RemoveContainer" containerID="c41e0c8ee6b9e1d84b96dc51ba09d436e2f5e86e3138bf78d4176b048eac4531" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.784438 4761 scope.go:117] "RemoveContainer" containerID="b0246e180b5baf24dd4b04c8a155233673df8f15aca70ad693f75b747959e68e" Nov 25 23:21:02 crc kubenswrapper[4761]: E1125 23:21:02.784849 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-66b769fbc8-xf6tp_openstack-operators(3daf70a9-4dab-4ace-82c0-9c7e48512cf6)\"" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.798232 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.800412 4761 generic.go:334] "Generic (PLEG): container finished" podID="bca3f822-b2ce-401e-b219-52d29e761c49" containerID="59e712783db882aec67f4f35b383f5f0eeea0c4a00738d65b051ae9c4ff4d247" exitCode=1 Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.800470 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerDied","Data":"59e712783db882aec67f4f35b383f5f0eeea0c4a00738d65b051ae9c4ff4d247"} Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.802423 4761 scope.go:117] "RemoveContainer" containerID="59e712783db882aec67f4f35b383f5f0eeea0c4a00738d65b051ae9c4ff4d247" Nov 25 23:21:02 crc kubenswrapper[4761]: E1125 23:21:02.802772 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-6d87477c79-vr8fq_metallb-system(bca3f822-b2ce-401e-b219-52d29e761c49)\"" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.816501 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.825929 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.869955 4761 scope.go:117] "RemoveContainer" containerID="8a4d4c2ff41784ed44785b9cd6da3eff1505acb493207e79ba8518b310a92527" Nov 25 23:21:02 crc kubenswrapper[4761]: I1125 23:21:02.942413 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-t4jrq" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.011752 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.026522 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.026848 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.054781 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.202522 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-cfkgt" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.252044 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.397992 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.427840 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.640079 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.675048 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.676400 4761 scope.go:117] "RemoveContainer" containerID="edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96" Nov 25 23:21:03 crc kubenswrapper[4761]: E1125 23:21:03.677000 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c5dd6c96-z8jr7_openstack-operators(2e34c701-6bbc-4000-a90a-aeca2333b69b)\"" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.823466 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 23:21:03 crc kubenswrapper[4761]: I1125 23:21:03.935547 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.050118 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.103803 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.179854 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.215061 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.215991 4761 scope.go:117] "RemoveContainer" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" Nov 25 23:21:04 crc kubenswrapper[4761]: E1125 23:21:04.216338 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.216380 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.268977 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.329099 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.399206 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.413069 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.566386 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.601149 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.722621 4761 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.820265 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.821203 4761 scope.go:117] "RemoveContainer" containerID="c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22" Nov 25 23:21:04 crc kubenswrapper[4761]: E1125 23:21:04.821572 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-7f54679d79-6ldpn_openstack-operators(bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b)\"" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.840943 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.841247 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.841501 4761 scope.go:117] "RemoveContainer" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" Nov 25 23:21:04 crc kubenswrapper[4761]: E1125 23:21:04.841779 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:21:04 crc kubenswrapper[4761]: I1125 23:21:04.896465 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.042477 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.054100 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.078477 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-k2qxk" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.098806 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.127474 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.147830 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-c9dlg" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.180491 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.315213 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.423450 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.487820 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.491982 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.709735 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.775992 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.839519 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.843470 4761 scope.go:117] "RemoveContainer" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" Nov 25 23:21:05 crc kubenswrapper[4761]: E1125 23:21:05.843910 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.849371 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 23:21:05 crc kubenswrapper[4761]: I1125 23:21:05.913737 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.087724 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.119941 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.149663 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.175216 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.208901 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.226117 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.302565 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.325548 4761 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.377816 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.396329 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-n6xdd" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.510191 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.614160 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.673553 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.685250 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.700379 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.755508 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.762728 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.795770 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.833295 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.836326 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.848632 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.871184 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-72m9f" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.990619 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 23:21:06 crc kubenswrapper[4761]: I1125 23:21:06.998917 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.055027 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.076858 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.118684 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kjg7n" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.144398 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.209956 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.284746 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.295304 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.315824 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.331974 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.332079 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.346164 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.380253 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.430754 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.443332 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.467659 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.533197 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.545067 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.565559 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.575599 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.614884 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.674830 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.716375 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.723378 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.744968 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.803007 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-fjnq8" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.829751 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.885665 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.910079 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.935546 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.938366 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.974340 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 23:21:07 crc kubenswrapper[4761]: I1125 23:21:07.984213 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.056082 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.063583 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-ch5fs" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.111344 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.114708 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.124253 4761 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.130761 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.146785 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.154689 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.172155 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.174620 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.201169 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-7mzz2" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.225138 4761 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.231160 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.308250 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.327912 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.486836 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.490412 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.499089 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.583438 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.583494 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.677660 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.700967 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.829384 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.839170 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.898445 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.903835 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.956567 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 23:21:08 crc kubenswrapper[4761]: I1125 23:21:08.991299 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.022128 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.024782 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.046447 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.137328 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.246216 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.253988 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.255632 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.273465 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.314086 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.316140 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.333908 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.356236 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.372499 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.415485 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.515867 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.529928 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-2tj2k" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.574999 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.604682 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.604797 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.605571 4761 scope.go:117] "RemoveContainer" containerID="b0246e180b5baf24dd4b04c8a155233673df8f15aca70ad693f75b747959e68e" Nov 25 23:21:09 crc kubenswrapper[4761]: E1125 23:21:09.605982 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-66b769fbc8-xf6tp_openstack-operators(3daf70a9-4dab-4ace-82c0-9c7e48512cf6)\"" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" podUID="3daf70a9-4dab-4ace-82c0-9c7e48512cf6" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.639170 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.665791 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.733234 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.792399 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.842114 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.843984 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.892073 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.899371 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-752bb" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.904685 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.920118 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.950105 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.975054 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.990517 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 23:21:09 crc kubenswrapper[4761]: I1125 23:21:09.997545 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.004726 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.005511 4761 scope.go:117] "RemoveContainer" containerID="59e712783db882aec67f4f35b383f5f0eeea0c4a00738d65b051ae9c4ff4d247" Nov 25 23:21:10 crc kubenswrapper[4761]: E1125 23:21:10.005886 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-6d87477c79-vr8fq_metallb-system(bca3f822-b2ce-401e-b219-52d29e761c49)\"" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" podUID="bca3f822-b2ce-401e-b219-52d29e761c49" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.011546 4761 scope.go:117] "RemoveContainer" containerID="82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8" Nov 25 23:21:10 crc kubenswrapper[4761]: E1125 23:21:10.011813 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-cskvg_openstack-operators(91b5b48e-a642-45e2-a84a-d449123e57bb)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" podUID="91b5b48e-a642-45e2-a84a-d449123e57bb" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.051526 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.058384 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.151102 4761 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.190942 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.227273 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.255072 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.287064 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.287148 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.287250 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.342900 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.455084 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.457644 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.491966 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.492541 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.494777 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.496292 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.525760 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.550938 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.559264 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.579087 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.799426 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.810643 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 23:21:10 crc kubenswrapper[4761]: I1125 23:21:10.986674 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.052531 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.053254 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.192979 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.202913 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.284451 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.318048 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.343536 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.423928 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.444373 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.446946 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.458807 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.516651 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-p5ttb" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.677995 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.760379 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.827059 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.837762 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 23:21:11 crc kubenswrapper[4761]: I1125 23:21:11.961929 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.001251 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.065021 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.093130 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.103537 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.328515 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.353746 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.402919 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.428789 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-5qvmg" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.489754 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.490422 4761 scope.go:117] "RemoveContainer" containerID="545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6" Nov 25 23:21:12 crc kubenswrapper[4761]: E1125 23:21:12.490690 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-6bbbcc9f68-dcp42_openstack-operators(0717f7d1-d5b3-408e-878e-10c62251448a)\"" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" podUID="0717f7d1-d5b3-408e-878e-10c62251448a" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.565315 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.600088 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.675181 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.691799 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.715826 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.770581 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.775165 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.787636 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.818349 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.825561 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.832100 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.877163 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.893737 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.919316 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.922115 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.928249 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 23:21:12 crc kubenswrapper[4761]: I1125 23:21:12.945207 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.012986 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.043906 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.189375 4761 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.189597 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09" gracePeriod=5 Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.195023 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.198854 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.249587 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.274886 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.405111 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.412094 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.458640 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.467951 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-msgsj" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.478842 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.650411 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.675162 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.676285 4761 scope.go:117] "RemoveContainer" containerID="edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96" Nov 25 23:21:13 crc kubenswrapper[4761]: E1125 23:21:13.676809 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c5dd6c96-z8jr7_openstack-operators(2e34c701-6bbc-4000-a90a-aeca2333b69b)\"" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" podUID="2e34c701-6bbc-4000-a90a-aeca2333b69b" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.712288 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.744341 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 23:21:13 crc kubenswrapper[4761]: I1125 23:21:13.931001 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.076633 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.097168 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.132962 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.214917 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.215499 4761 scope.go:117] "RemoveContainer" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" Nov 25 23:21:14 crc kubenswrapper[4761]: E1125 23:21:14.215715 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6c6d6c68d5-v5cfc_openstack-operators(8aba4458-8023-4b9f-8504-6fab49f5250c)\"" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" podUID="8aba4458-8023-4b9f-8504-6fab49f5250c" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.271331 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.396467 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.415917 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-pxsr2" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.451079 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-76m6w" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.505158 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.527436 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.544139 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.556954 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.560680 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-g4w2q" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.594406 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.641212 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.656008 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.698167 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.730756 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.800437 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-mfh54" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.821040 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.822164 4761 scope.go:117] "RemoveContainer" containerID="c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22" Nov 25 23:21:14 crc kubenswrapper[4761]: E1125 23:21:14.822391 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-7f54679d79-6ldpn_openstack-operators(bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b)\"" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" podUID="bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.870202 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.908992 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 23:21:14 crc kubenswrapper[4761]: I1125 23:21:14.929772 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.057421 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.059033 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.128571 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.141861 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.148262 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.169924 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.171924 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:21:15 crc kubenswrapper[4761]: E1125 23:21:15.172230 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" containerName="installer" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.172252 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" containerName="installer" Nov 25 23:21:15 crc kubenswrapper[4761]: E1125 23:21:15.172269 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.172279 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.172436 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.172452 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="6568a4b2-1689-4cfa-84a7-1d952014beac" containerName="installer" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.173597 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.198330 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.229094 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.282107 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.316235 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.316430 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.316563 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6g58\" (UniqueName: \"kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.412644 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.418580 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6g58\" (UniqueName: \"kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.418800 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.418987 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.419310 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.419474 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.427664 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.434899 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.443222 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6g58\" (UniqueName: \"kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58\") pod \"redhat-operators-g8n8s\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.509815 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.516517 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.577771 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.578971 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.595592 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.625046 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.648657 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.734429 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.750280 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.844984 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.920594 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 23:21:15 crc kubenswrapper[4761]: I1125 23:21:15.976476 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:21:15 crc kubenswrapper[4761]: W1125 23:21:15.983051 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ab37abb_3734_45cb_88b3_fa726716d5e0.slice/crio-96b8dc246b7daffb6720635fd961188f9b73f1ddd5a3e910dfbb1d6e13a3c5a7 WatchSource:0}: Error finding container 96b8dc246b7daffb6720635fd961188f9b73f1ddd5a3e910dfbb1d6e13a3c5a7: Status 404 returned error can't find the container with id 96b8dc246b7daffb6720635fd961188f9b73f1ddd5a3e910dfbb1d6e13a3c5a7 Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.190012 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pvqmg" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.285873 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-txgjr" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.371297 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.491310 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.495549 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.511476 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.638494 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.705023 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.891750 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.938423 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.959981 4761 generic.go:334] "Generic (PLEG): container finished" podID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerID="7a424506ad005f294b44d33035f3d4fb2a1c0b568e46ce474abbef45903f2f19" exitCode=0 Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.960049 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerDied","Data":"7a424506ad005f294b44d33035f3d4fb2a1c0b568e46ce474abbef45903f2f19"} Nov 25 23:21:16 crc kubenswrapper[4761]: I1125 23:21:16.960088 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerStarted","Data":"96b8dc246b7daffb6720635fd961188f9b73f1ddd5a3e910dfbb1d6e13a3c5a7"} Nov 25 23:21:17 crc kubenswrapper[4761]: I1125 23:21:17.135213 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 23:21:17 crc kubenswrapper[4761]: I1125 23:21:17.209821 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-x4c9s" Nov 25 23:21:17 crc kubenswrapper[4761]: I1125 23:21:17.214152 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.011513 4761 scope.go:117] "RemoveContainer" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" Nov 25 23:21:18 crc kubenswrapper[4761]: E1125 23:21:18.011800 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-8559b887bb-frnqv_openstack-operators(47a4708b-be53-4783-b9c1-9536d90b9e4c)\"" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" podUID="47a4708b-be53-4783-b9c1-9536d90b9e4c" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.588583 4761 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.775457 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.775531 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.874829 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.874878 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.874906 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.874931 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.874957 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875020 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875054 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875067 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875209 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875594 4761 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875613 4761 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875624 4761 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.875636 4761 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.882822 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.976538 4761 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.977902 4761 generic.go:334] "Generic (PLEG): container finished" podID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerID="3c034ac3e26e672c8994902535c9d35851126ea915d738f47f58f0ddaadd3b14" exitCode=0 Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.977987 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerDied","Data":"3c034ac3e26e672c8994902535c9d35851126ea915d738f47f58f0ddaadd3b14"} Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.979782 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.979844 4761 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09" exitCode=137 Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.979890 4761 scope.go:117] "RemoveContainer" containerID="83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09" Nov 25 23:21:18 crc kubenswrapper[4761]: I1125 23:21:18.979917 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 23:21:19 crc kubenswrapper[4761]: I1125 23:21:19.013751 4761 scope.go:117] "RemoveContainer" containerID="83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09" Nov 25 23:21:19 crc kubenswrapper[4761]: E1125 23:21:19.014485 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09\": container with ID starting with 83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09 not found: ID does not exist" containerID="83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09" Nov 25 23:21:19 crc kubenswrapper[4761]: I1125 23:21:19.014567 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09"} err="failed to get container status \"83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09\": rpc error: code = NotFound desc = could not find container \"83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09\": container with ID starting with 83b98970fbc05439e878800b20c021baee3f3eb73a0d1e7071646b09b3fcfc09 not found: ID does not exist" Nov 25 23:21:19 crc kubenswrapper[4761]: I1125 23:21:19.021876 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 23:21:19 crc kubenswrapper[4761]: I1125 23:21:19.993725 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerStarted","Data":"ce2d671d2deea5e40e642f830f72e4d243ff4b8d0e9a96ee345ab2340929f1a4"} Nov 25 23:21:20 crc kubenswrapper[4761]: I1125 23:21:20.028759 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8n8s" podStartSLOduration=8.546988333 podStartE2EDuration="11.028682955s" podCreationTimestamp="2025-11-25 23:21:09 +0000 UTC" firstStartedPulling="2025-11-25 23:21:16.964866863 +0000 UTC m=+1472.698212738" lastFinishedPulling="2025-11-25 23:21:19.446561485 +0000 UTC m=+1475.179907360" observedRunningTime="2025-11-25 23:21:20.017112185 +0000 UTC m=+1475.750458060" watchObservedRunningTime="2025-11-25 23:21:20.028682955 +0000 UTC m=+1475.762028820" Nov 25 23:21:21 crc kubenswrapper[4761]: I1125 23:21:21.128046 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:21:21 crc kubenswrapper[4761]: I1125 23:21:21.128138 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:21:23 crc kubenswrapper[4761]: I1125 23:21:23.011834 4761 scope.go:117] "RemoveContainer" containerID="545ef4a95751fdc889d04fdadb3777e7490bab77672309778aaaa8bb56193fe6" Nov 25 23:21:24 crc kubenswrapper[4761]: I1125 23:21:24.011206 4761 scope.go:117] "RemoveContainer" containerID="b0246e180b5baf24dd4b04c8a155233673df8f15aca70ad693f75b747959e68e" Nov 25 23:21:24 crc kubenswrapper[4761]: I1125 23:21:24.011514 4761 scope.go:117] "RemoveContainer" containerID="59e712783db882aec67f4f35b383f5f0eeea0c4a00738d65b051ae9c4ff4d247" Nov 25 23:21:24 crc kubenswrapper[4761]: I1125 23:21:24.056553 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" event={"ID":"0717f7d1-d5b3-408e-878e-10c62251448a","Type":"ContainerStarted","Data":"ffc2b00fd52fc0dfac8aef4893d45f9e734d87ee1eb5f406c0006c2375d913e3"} Nov 25 23:21:24 crc kubenswrapper[4761]: I1125 23:21:24.056845 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.015929 4761 scope.go:117] "RemoveContainer" containerID="82dfd26ce7a02fa08706f02b5fdfd9c51dac77d4cc2ab0e1384ae84dde135bd8" Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.075625 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" event={"ID":"bca3f822-b2ce-401e-b219-52d29e761c49","Type":"ContainerStarted","Data":"58f0f9edcf44cbf076e79c998b90fc8d572650a7132b988560fb3690b49bd517"} Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.076589 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.079094 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" event={"ID":"3daf70a9-4dab-4ace-82c0-9c7e48512cf6","Type":"ContainerStarted","Data":"2ce950044828562f6ce66823a246682c44dd2e967bc0cf4c8a7262403a7e8cc0"} Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.079495 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.510626 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:25 crc kubenswrapper[4761]: I1125 23:21:25.510687 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:26 crc kubenswrapper[4761]: I1125 23:21:26.012761 4761 scope.go:117] "RemoveContainer" containerID="89c4380d0a255722c0f5a219b263b989560f69e46161242c35c1a5c9449211ce" Nov 25 23:21:26 crc kubenswrapper[4761]: I1125 23:21:26.092680 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cskvg" event={"ID":"91b5b48e-a642-45e2-a84a-d449123e57bb","Type":"ContainerStarted","Data":"5351865e1f41bb91b0ceb4eb97888e0995b06687fcbd3aafe9faa8a8f3271c22"} Nov 25 23:21:26 crc kubenswrapper[4761]: I1125 23:21:26.585376 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8n8s" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="registry-server" probeResult="failure" output=< Nov 25 23:21:26 crc kubenswrapper[4761]: timeout: failed to connect service ":50051" within 1s Nov 25 23:21:26 crc kubenswrapper[4761]: > Nov 25 23:21:27 crc kubenswrapper[4761]: I1125 23:21:27.011056 4761 scope.go:117] "RemoveContainer" containerID="edd6c44d18559be69bc249597585a857b03b1b52a21386568dc7426344d81b96" Nov 25 23:21:27 crc kubenswrapper[4761]: I1125 23:21:27.109173 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" event={"ID":"8aba4458-8023-4b9f-8504-6fab49f5250c","Type":"ContainerStarted","Data":"ba1aac9a42a0a7ab4cdf8de445c9f57e99e1e730b587085c1e4e5aad2446b83f"} Nov 25 23:21:27 crc kubenswrapper[4761]: I1125 23:21:27.109414 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:21:28 crc kubenswrapper[4761]: I1125 23:21:28.134265 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" event={"ID":"2e34c701-6bbc-4000-a90a-aeca2333b69b","Type":"ContainerStarted","Data":"7f3a8d9544a52b65d2ca09bd0160aeb323e3d27168df032e24c1473c3043aeb5"} Nov 25 23:21:28 crc kubenswrapper[4761]: I1125 23:21:28.134769 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:21:29 crc kubenswrapper[4761]: I1125 23:21:29.617481 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-66b769fbc8-xf6tp" Nov 25 23:21:30 crc kubenswrapper[4761]: I1125 23:21:30.011680 4761 scope.go:117] "RemoveContainer" containerID="c29ec5a458aa88fae3f3922d469d34e049c1e678159fcb390d58a41ed6addd22" Nov 25 23:21:30 crc kubenswrapper[4761]: I1125 23:21:30.012363 4761 scope.go:117] "RemoveContainer" containerID="30296285a193f01ee470f8f75f7a607855a0e189f243e55f85aaf1ae4870ee4f" Nov 25 23:21:31 crc kubenswrapper[4761]: I1125 23:21:31.167559 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" event={"ID":"47a4708b-be53-4783-b9c1-9536d90b9e4c","Type":"ContainerStarted","Data":"af172ca35c7a93e6a84d5fab4f6be9825ec9d1521fa359849ac25cf7e3131e63"} Nov 25 23:21:31 crc kubenswrapper[4761]: I1125 23:21:31.168176 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:21:31 crc kubenswrapper[4761]: I1125 23:21:31.170337 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" event={"ID":"bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b","Type":"ContainerStarted","Data":"087d6ecc0ae4a089c76227237b4c0eefff4620bcbc88e0d3fe5b0b8b7fe09934"} Nov 25 23:21:31 crc kubenswrapper[4761]: I1125 23:21:31.170581 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:21:32 crc kubenswrapper[4761]: I1125 23:21:32.496458 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6bbbcc9f68-dcp42" Nov 25 23:21:33 crc kubenswrapper[4761]: I1125 23:21:33.680793 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c5dd6c96-z8jr7" Nov 25 23:21:34 crc kubenswrapper[4761]: I1125 23:21:34.221413 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6c6d6c68d5-v5cfc" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.335913 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.337590 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.355917 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.457136 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5lt2\" (UniqueName: \"kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.457204 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.457283 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.544244 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.546013 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.551870 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.558551 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.558568 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5lt2\" (UniqueName: \"kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.558646 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.558728 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.559384 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.559484 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.589814 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5lt2\" (UniqueName: \"kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2\") pod \"certified-operators-2lnnb\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.627171 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.652984 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.660364 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.660968 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.661091 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9jjq\" (UniqueName: \"kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.762455 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.762539 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9jjq\" (UniqueName: \"kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.762603 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.763757 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.763907 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.825634 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9jjq\" (UniqueName: \"kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq\") pod \"community-operators-m6f54\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:35 crc kubenswrapper[4761]: I1125 23:21:35.865671 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:36 crc kubenswrapper[4761]: W1125 23:21:36.175890 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc9672bd_1e5e_4ef7_8713_145c1a8a6cd6.slice/crio-ac6d9247a51bbd2cb0aec142c1876c604cfcd54e8eccb104ec01e530741dda16 WatchSource:0}: Error finding container ac6d9247a51bbd2cb0aec142c1876c604cfcd54e8eccb104ec01e530741dda16: Status 404 returned error can't find the container with id ac6d9247a51bbd2cb0aec142c1876c604cfcd54e8eccb104ec01e530741dda16 Nov 25 23:21:36 crc kubenswrapper[4761]: I1125 23:21:36.178614 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:21:36 crc kubenswrapper[4761]: I1125 23:21:36.223049 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerStarted","Data":"ac6d9247a51bbd2cb0aec142c1876c604cfcd54e8eccb104ec01e530741dda16"} Nov 25 23:21:36 crc kubenswrapper[4761]: I1125 23:21:36.313541 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:21:36 crc kubenswrapper[4761]: W1125 23:21:36.326800 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod933d00e9_8ab4_4a1b_a7d1_ae18bf7e2a76.slice/crio-2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69 WatchSource:0}: Error finding container 2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69: Status 404 returned error can't find the container with id 2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69 Nov 25 23:21:37 crc kubenswrapper[4761]: I1125 23:21:37.242508 4761 generic.go:334] "Generic (PLEG): container finished" podID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerID="079d882b2853b95d26ff0c45ae5e6a71e60a6d34a212f6a5fe1684be942deaf4" exitCode=0 Nov 25 23:21:37 crc kubenswrapper[4761]: I1125 23:21:37.242592 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerDied","Data":"079d882b2853b95d26ff0c45ae5e6a71e60a6d34a212f6a5fe1684be942deaf4"} Nov 25 23:21:37 crc kubenswrapper[4761]: I1125 23:21:37.243040 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerStarted","Data":"2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69"} Nov 25 23:21:37 crc kubenswrapper[4761]: I1125 23:21:37.245823 4761 generic.go:334] "Generic (PLEG): container finished" podID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerID="47da488f0e61b2d2570914fcde745b21f14dc30296e2a43067e868b69eca66e3" exitCode=0 Nov 25 23:21:37 crc kubenswrapper[4761]: I1125 23:21:37.245890 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerDied","Data":"47da488f0e61b2d2570914fcde745b21f14dc30296e2a43067e868b69eca66e3"} Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.262738 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerStarted","Data":"298e4a84d87ae69e3d97223d19c137c890fe5774d3fbda61593edf3768988a24"} Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.341096 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.342444 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.352559 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.522126 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.522201 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj2bx\" (UniqueName: \"kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.522251 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.624056 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj2bx\" (UniqueName: \"kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.624146 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.624191 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.624628 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.625098 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.670527 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj2bx\" (UniqueName: \"kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx\") pod \"redhat-operators-4c4l4\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:38 crc kubenswrapper[4761]: I1125 23:21:38.729043 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.232091 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:21:39 crc kubenswrapper[4761]: W1125 23:21:39.241306 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod032c8467_d8b9_4345_88a7_662e363a3174.slice/crio-677de3c35dd665417de67eabc0ebc40d6a67d1d9841242c0647231f92e4ec47a WatchSource:0}: Error finding container 677de3c35dd665417de67eabc0ebc40d6a67d1d9841242c0647231f92e4ec47a: Status 404 returned error can't find the container with id 677de3c35dd665417de67eabc0ebc40d6a67d1d9841242c0647231f92e4ec47a Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.274361 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerStarted","Data":"677de3c35dd665417de67eabc0ebc40d6a67d1d9841242c0647231f92e4ec47a"} Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.278273 4761 generic.go:334] "Generic (PLEG): container finished" podID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerID="a9f8181244335bf706b8241ac2d5b4781c9d981d199747dcccfbcd1af841b226" exitCode=0 Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.278326 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerDied","Data":"a9f8181244335bf706b8241ac2d5b4781c9d981d199747dcccfbcd1af841b226"} Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.293019 4761 generic.go:334] "Generic (PLEG): container finished" podID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerID="298e4a84d87ae69e3d97223d19c137c890fe5774d3fbda61593edf3768988a24" exitCode=0 Nov 25 23:21:39 crc kubenswrapper[4761]: I1125 23:21:39.293096 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerDied","Data":"298e4a84d87ae69e3d97223d19c137c890fe5774d3fbda61593edf3768988a24"} Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.302059 4761 generic.go:334] "Generic (PLEG): container finished" podID="032c8467-d8b9-4345-88a7-662e363a3174" containerID="fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140" exitCode=0 Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.302154 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerDied","Data":"fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140"} Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.305954 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerStarted","Data":"390337cc4cae0d91372d078a75e6fe818bf939c5f51c823ec46ae84d2e1618e7"} Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.309186 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerStarted","Data":"b1d0bbd28350eb43033ecd8d25e322b4bb2832068b86db5c049fcfacc318b65a"} Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.350439 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m6f54" podStartSLOduration=2.897427738 podStartE2EDuration="5.350416562s" podCreationTimestamp="2025-11-25 23:21:35 +0000 UTC" firstStartedPulling="2025-11-25 23:21:37.245959112 +0000 UTC m=+1492.979304977" lastFinishedPulling="2025-11-25 23:21:39.698947926 +0000 UTC m=+1495.432293801" observedRunningTime="2025-11-25 23:21:40.342988914 +0000 UTC m=+1496.076334779" watchObservedRunningTime="2025-11-25 23:21:40.350416562 +0000 UTC m=+1496.083762417" Nov 25 23:21:40 crc kubenswrapper[4761]: I1125 23:21:40.369716 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2lnnb" podStartSLOduration=2.905764562 podStartE2EDuration="5.369676258s" podCreationTimestamp="2025-11-25 23:21:35 +0000 UTC" firstStartedPulling="2025-11-25 23:21:37.248031187 +0000 UTC m=+1492.981377062" lastFinishedPulling="2025-11-25 23:21:39.711942913 +0000 UTC m=+1495.445288758" observedRunningTime="2025-11-25 23:21:40.364324845 +0000 UTC m=+1496.097670710" watchObservedRunningTime="2025-11-25 23:21:40.369676258 +0000 UTC m=+1496.103022103" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.318585 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerStarted","Data":"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859"} Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.745797 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.747324 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.755027 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.874053 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnntl\" (UniqueName: \"kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.874248 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.874301 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.975614 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnntl\" (UniqueName: \"kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.975757 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.975795 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.976428 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:41 crc kubenswrapper[4761]: I1125 23:21:41.976587 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:42 crc kubenswrapper[4761]: I1125 23:21:42.003577 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnntl\" (UniqueName: \"kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl\") pod \"redhat-operators-fwb7t\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:42 crc kubenswrapper[4761]: I1125 23:21:42.074249 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:42 crc kubenswrapper[4761]: I1125 23:21:42.339377 4761 generic.go:334] "Generic (PLEG): container finished" podID="032c8467-d8b9-4345-88a7-662e363a3174" containerID="b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859" exitCode=0 Nov 25 23:21:42 crc kubenswrapper[4761]: I1125 23:21:42.339419 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerDied","Data":"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859"} Nov 25 23:21:42 crc kubenswrapper[4761]: I1125 23:21:42.590023 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.142399 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.144311 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.159989 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.213497 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kscjj\" (UniqueName: \"kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.213676 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.213921 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.315159 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.315318 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.315442 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kscjj\" (UniqueName: \"kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.315821 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.316031 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.347384 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kscjj\" (UniqueName: \"kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj\") pod \"redhat-operators-vb27h\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.356985 4761 generic.go:334] "Generic (PLEG): container finished" podID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerID="ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48" exitCode=0 Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.357092 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerDied","Data":"ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48"} Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.357135 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerStarted","Data":"8e512a5babf65e1ebc7425bbd57a02e4ce8103e38d4b01181277ac17398b30fe"} Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.361960 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerStarted","Data":"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6"} Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.413664 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4c4l4" podStartSLOduration=2.917840779 podStartE2EDuration="5.413644519s" podCreationTimestamp="2025-11-25 23:21:38 +0000 UTC" firstStartedPulling="2025-11-25 23:21:40.303914517 +0000 UTC m=+1496.037260362" lastFinishedPulling="2025-11-25 23:21:42.799718267 +0000 UTC m=+1498.533064102" observedRunningTime="2025-11-25 23:21:43.405164972 +0000 UTC m=+1499.138510827" watchObservedRunningTime="2025-11-25 23:21:43.413644519 +0000 UTC m=+1499.146990364" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.502709 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:43 crc kubenswrapper[4761]: I1125 23:21:43.977756 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.050053 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hdrz6"] Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.057128 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8"] Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.063029 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hdrz6"] Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.069378 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-edb2-account-create-update-cn8n8"] Nov 25 23:21:44 crc kubenswrapper[4761]: W1125 23:21:44.090099 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33daa866_c097_4020_9432_0b76875fef0c.slice/crio-a6aa6f892a7dca46bfd360026dc3955f7be5a700577aa98b3d3fea10813779fd WatchSource:0}: Error finding container a6aa6f892a7dca46bfd360026dc3955f7be5a700577aa98b3d3fea10813779fd: Status 404 returned error can't find the container with id a6aa6f892a7dca46bfd360026dc3955f7be5a700577aa98b3d3fea10813779fd Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.368453 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerStarted","Data":"a6aa6f892a7dca46bfd360026dc3955f7be5a700577aa98b3d3fea10813779fd"} Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.826383 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7f54679d79-6ldpn" Nov 25 23:21:44 crc kubenswrapper[4761]: I1125 23:21:44.854652 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-8559b887bb-frnqv" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.050163 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d39ec9-e2eb-41a4-83d8-f0f84fb3681f" path="/var/lib/kubelet/pods/32d39ec9-e2eb-41a4-83d8-f0f84fb3681f/volumes" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.053891 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cda8b5ee-b6f4-4eb2-a095-f9647da330f6" path="/var/lib/kubelet/pods/cda8b5ee-b6f4-4eb2-a095-f9647da330f6/volumes" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.376813 4761 generic.go:334] "Generic (PLEG): container finished" podID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerID="b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc" exitCode=0 Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.376894 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerDied","Data":"b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc"} Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.379520 4761 generic.go:334] "Generic (PLEG): container finished" podID="33daa866-c097-4020-9432-0b76875fef0c" containerID="0777d443f53dc2de9bc9b35674299a8ca4ac91c306022dc17d166e468373ba13" exitCode=0 Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.379605 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerDied","Data":"0777d443f53dc2de9bc9b35674299a8ca4ac91c306022dc17d166e468373ba13"} Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.653183 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.655156 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.736613 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.867279 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.867327 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:45 crc kubenswrapper[4761]: I1125 23:21:45.928801 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:46 crc kubenswrapper[4761]: I1125 23:21:46.387170 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerStarted","Data":"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a"} Nov 25 23:21:46 crc kubenswrapper[4761]: I1125 23:21:46.391124 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerStarted","Data":"13fbe95f6c4583ee40abe0c599356ca54b4766716b57757a107f5f06d445990b"} Nov 25 23:21:46 crc kubenswrapper[4761]: I1125 23:21:46.407410 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fwb7t" podStartSLOduration=2.995836631 podStartE2EDuration="5.407394355s" podCreationTimestamp="2025-11-25 23:21:41 +0000 UTC" firstStartedPulling="2025-11-25 23:21:43.359228622 +0000 UTC m=+1499.092574467" lastFinishedPulling="2025-11-25 23:21:45.770786346 +0000 UTC m=+1501.504132191" observedRunningTime="2025-11-25 23:21:46.40422108 +0000 UTC m=+1502.137566935" watchObservedRunningTime="2025-11-25 23:21:46.407394355 +0000 UTC m=+1502.140740190" Nov 25 23:21:46 crc kubenswrapper[4761]: I1125 23:21:46.451003 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:21:46 crc kubenswrapper[4761]: I1125 23:21:46.466941 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.157954 4761 scope.go:117] "RemoveContainer" containerID="e5eeb6a02324099e65ff127469671ffbc53b2b026e5594dd23f3b394d041beeb" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.174030 4761 scope.go:117] "RemoveContainer" containerID="9ceee44d6828c43af4df267956dcd49e8cddaca7d565adff59bc4070c472a393" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.210267 4761 scope.go:117] "RemoveContainer" containerID="74b2da04d0c1f7778e6fcbc66f71cef9e69b598da0ef2ee653d17a1e389105fd" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.235874 4761 scope.go:117] "RemoveContainer" containerID="8d08c0aa3234212faeeee9a61a65a81bef37e77cdfc35ae022b7b64b50213129" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.275618 4761 scope.go:117] "RemoveContainer" containerID="a3541be67296fdb6f307d37a883e8d6326768c9d2d067c1fd4fa009f7a7acf31" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.297107 4761 scope.go:117] "RemoveContainer" containerID="b7d0a09b081655af53b4e8ded2c2d3bb8e3539ff5beea727ee082a98215c4436" Nov 25 23:21:47 crc kubenswrapper[4761]: I1125 23:21:47.322204 4761 scope.go:117] "RemoveContainer" containerID="f4d902e8a57f9b519513f37a8f3fa687f603d8cb2a868c355c857cc73dae40c0" Nov 25 23:21:48 crc kubenswrapper[4761]: I1125 23:21:48.418228 4761 generic.go:334] "Generic (PLEG): container finished" podID="33daa866-c097-4020-9432-0b76875fef0c" containerID="13fbe95f6c4583ee40abe0c599356ca54b4766716b57757a107f5f06d445990b" exitCode=0 Nov 25 23:21:48 crc kubenswrapper[4761]: I1125 23:21:48.418304 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerDied","Data":"13fbe95f6c4583ee40abe0c599356ca54b4766716b57757a107f5f06d445990b"} Nov 25 23:21:48 crc kubenswrapper[4761]: I1125 23:21:48.729865 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:48 crc kubenswrapper[4761]: I1125 23:21:48.730187 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:48 crc kubenswrapper[4761]: I1125 23:21:48.780155 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:49 crc kubenswrapper[4761]: I1125 23:21:49.429364 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerStarted","Data":"42c97cc799995e28f035584a03b9388ebaf0e07942708477ecae89ddf115bed5"} Nov 25 23:21:49 crc kubenswrapper[4761]: I1125 23:21:49.466877 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vb27h" podStartSLOduration=2.958873084 podStartE2EDuration="6.466859691s" podCreationTimestamp="2025-11-25 23:21:43 +0000 UTC" firstStartedPulling="2025-11-25 23:21:45.380533175 +0000 UTC m=+1501.113879010" lastFinishedPulling="2025-11-25 23:21:48.888519772 +0000 UTC m=+1504.621865617" observedRunningTime="2025-11-25 23:21:49.45936516 +0000 UTC m=+1505.192711025" watchObservedRunningTime="2025-11-25 23:21:49.466859691 +0000 UTC m=+1505.200205536" Nov 25 23:21:49 crc kubenswrapper[4761]: I1125 23:21:49.485207 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:21:51 crc kubenswrapper[4761]: I1125 23:21:51.128376 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:21:51 crc kubenswrapper[4761]: I1125 23:21:51.128729 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:21:52 crc kubenswrapper[4761]: I1125 23:21:52.074496 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:52 crc kubenswrapper[4761]: I1125 23:21:52.074552 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:52 crc kubenswrapper[4761]: I1125 23:21:52.123429 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:52 crc kubenswrapper[4761]: I1125 23:21:52.496907 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:21:53 crc kubenswrapper[4761]: I1125 23:21:53.503741 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:53 crc kubenswrapper[4761]: I1125 23:21:53.504124 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:21:54 crc kubenswrapper[4761]: I1125 23:21:54.573524 4761 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vb27h" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="registry-server" probeResult="failure" output=< Nov 25 23:21:54 crc kubenswrapper[4761]: timeout: failed to connect service ":50051" within 1s Nov 25 23:21:54 crc kubenswrapper[4761]: > Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.010136 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d87477c79-vr8fq" Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.069742 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6kdq"] Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.083160 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6kdq"] Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.343592 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.344033 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2lnnb" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="registry-server" containerID="cri-o://b1d0bbd28350eb43033ecd8d25e322b4bb2832068b86db5c049fcfacc318b65a" gracePeriod=2 Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.527584 4761 generic.go:334] "Generic (PLEG): container finished" podID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerID="b1d0bbd28350eb43033ecd8d25e322b4bb2832068b86db5c049fcfacc318b65a" exitCode=0 Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.527620 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerDied","Data":"b1d0bbd28350eb43033ecd8d25e322b4bb2832068b86db5c049fcfacc318b65a"} Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.867072 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.934603 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities\") pod \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.934972 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content\") pod \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.935030 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5lt2\" (UniqueName: \"kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2\") pod \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\" (UID: \"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6\") " Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.935768 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities" (OuterVolumeSpecName: "utilities") pod "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" (UID: "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.940787 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2" (OuterVolumeSpecName: "kube-api-access-v5lt2") pod "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" (UID: "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6"). InnerVolumeSpecName "kube-api-access-v5lt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:00 crc kubenswrapper[4761]: I1125 23:22:00.994914 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" (UID: "bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.024221 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30de3be7-530b-4f0b-910b-eb1060028d38" path="/var/lib/kubelet/pods/30de3be7-530b-4f0b-910b-eb1060028d38/volumes" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.036587 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.036615 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.036625 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5lt2\" (UniqueName: \"kubernetes.io/projected/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6-kube-api-access-v5lt2\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.541631 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lnnb" event={"ID":"bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6","Type":"ContainerDied","Data":"ac6d9247a51bbd2cb0aec142c1876c604cfcd54e8eccb104ec01e530741dda16"} Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.541711 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lnnb" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.541753 4761 scope.go:117] "RemoveContainer" containerID="b1d0bbd28350eb43033ecd8d25e322b4bb2832068b86db5c049fcfacc318b65a" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.569463 4761 scope.go:117] "RemoveContainer" containerID="298e4a84d87ae69e3d97223d19c137c890fe5774d3fbda61593edf3768988a24" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.584893 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.596664 4761 scope.go:117] "RemoveContainer" containerID="47da488f0e61b2d2570914fcde745b21f14dc30296e2a43067e868b69eca66e3" Nov 25 23:22:01 crc kubenswrapper[4761]: I1125 23:22:01.597002 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2lnnb"] Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.024832 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" path="/var/lib/kubelet/pods/bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6/volumes" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.146117 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.146772 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m6f54" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="registry-server" containerID="cri-o://390337cc4cae0d91372d078a75e6fe818bf939c5f51c823ec46ae84d2e1618e7" gracePeriod=2 Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.552355 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.578851 4761 generic.go:334] "Generic (PLEG): container finished" podID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerID="390337cc4cae0d91372d078a75e6fe818bf939c5f51c823ec46ae84d2e1618e7" exitCode=0 Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.578929 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerDied","Data":"390337cc4cae0d91372d078a75e6fe818bf939c5f51c823ec46ae84d2e1618e7"} Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.578965 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6f54" event={"ID":"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76","Type":"ContainerDied","Data":"2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69"} Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.578980 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2afe5e2f41c1536ab6c9c74af7b336f5fa5b3388bb5d59d9ea9df578221e0a69" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.614612 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.624613 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.787109 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities\") pod \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.787204 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content\") pod \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.787298 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9jjq\" (UniqueName: \"kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq\") pod \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\" (UID: \"933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76\") " Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.788800 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities" (OuterVolumeSpecName: "utilities") pod "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" (UID: "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.794363 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq" (OuterVolumeSpecName: "kube-api-access-q9jjq") pod "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" (UID: "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76"). InnerVolumeSpecName "kube-api-access-q9jjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.835744 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" (UID: "933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.890197 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.890238 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:03 crc kubenswrapper[4761]: I1125 23:22:03.890257 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9jjq\" (UniqueName: \"kubernetes.io/projected/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76-kube-api-access-q9jjq\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:04 crc kubenswrapper[4761]: I1125 23:22:04.587058 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6f54" Nov 25 23:22:04 crc kubenswrapper[4761]: I1125 23:22:04.630365 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:22:04 crc kubenswrapper[4761]: I1125 23:22:04.638473 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m6f54"] Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.026880 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" path="/var/lib/kubelet/pods/933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76/volumes" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.140512 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.140938 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dtxht" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="registry-server" containerID="cri-o://960bb5300b34104f250886cdbc3862a2dd991ea88b658211150c1e5819545b24" gracePeriod=2 Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.596745 4761 generic.go:334] "Generic (PLEG): container finished" podID="346df741-cf07-4002-a8b6-5069c015aee3" containerID="960bb5300b34104f250886cdbc3862a2dd991ea88b658211150c1e5819545b24" exitCode=0 Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.596861 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerDied","Data":"960bb5300b34104f250886cdbc3862a2dd991ea88b658211150c1e5819545b24"} Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.597106 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtxht" event={"ID":"346df741-cf07-4002-a8b6-5069c015aee3","Type":"ContainerDied","Data":"d9ffda992202ddfb7687df31ddd1d88513d106f1270ec7e81cee31de28e8e083"} Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.597122 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9ffda992202ddfb7687df31ddd1d88513d106f1270ec7e81cee31de28e8e083" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.639923 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.821958 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content\") pod \"346df741-cf07-4002-a8b6-5069c015aee3\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.822467 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities\") pod \"346df741-cf07-4002-a8b6-5069c015aee3\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.822778 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhtg5\" (UniqueName: \"kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5\") pod \"346df741-cf07-4002-a8b6-5069c015aee3\" (UID: \"346df741-cf07-4002-a8b6-5069c015aee3\") " Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.823970 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities" (OuterVolumeSpecName: "utilities") pod "346df741-cf07-4002-a8b6-5069c015aee3" (UID: "346df741-cf07-4002-a8b6-5069c015aee3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.832807 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5" (OuterVolumeSpecName: "kube-api-access-dhtg5") pod "346df741-cf07-4002-a8b6-5069c015aee3" (UID: "346df741-cf07-4002-a8b6-5069c015aee3"). InnerVolumeSpecName "kube-api-access-dhtg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.849489 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "346df741-cf07-4002-a8b6-5069c015aee3" (UID: "346df741-cf07-4002-a8b6-5069c015aee3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.927129 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.927198 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhtg5\" (UniqueName: \"kubernetes.io/projected/346df741-cf07-4002-a8b6-5069c015aee3-kube-api-access-dhtg5\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:05 crc kubenswrapper[4761]: I1125 23:22:05.927226 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/346df741-cf07-4002-a8b6-5069c015aee3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:06 crc kubenswrapper[4761]: I1125 23:22:06.607437 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtxht" Nov 25 23:22:06 crc kubenswrapper[4761]: I1125 23:22:06.662449 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:22:06 crc kubenswrapper[4761]: I1125 23:22:06.679386 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtxht"] Nov 25 23:22:07 crc kubenswrapper[4761]: I1125 23:22:07.025938 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="346df741-cf07-4002-a8b6-5069c015aee3" path="/var/lib/kubelet/pods/346df741-cf07-4002-a8b6-5069c015aee3/volumes" Nov 25 23:22:08 crc kubenswrapper[4761]: I1125 23:22:08.033363 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-prggk"] Nov 25 23:22:08 crc kubenswrapper[4761]: I1125 23:22:08.043132 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-prggk"] Nov 25 23:22:09 crc kubenswrapper[4761]: I1125 23:22:09.028156 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4fd046a-5f86-4226-a001-c6a166b2fbad" path="/var/lib/kubelet/pods/a4fd046a-5f86-4226-a001-c6a166b2fbad/volumes" Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.338751 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.339562 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vb27h" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="registry-server" containerID="cri-o://42c97cc799995e28f035584a03b9388ebaf0e07942708477ecae89ddf115bed5" gracePeriod=2 Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.730161 4761 generic.go:334] "Generic (PLEG): container finished" podID="33daa866-c097-4020-9432-0b76875fef0c" containerID="42c97cc799995e28f035584a03b9388ebaf0e07942708477ecae89ddf115bed5" exitCode=0 Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.730395 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerDied","Data":"42c97cc799995e28f035584a03b9388ebaf0e07942708477ecae89ddf115bed5"} Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.935130 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.935710 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fwb7t" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="registry-server" containerID="cri-o://0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a" gracePeriod=2 Nov 25 23:22:18 crc kubenswrapper[4761]: I1125 23:22:18.982297 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.066503 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content\") pod \"33daa866-c097-4020-9432-0b76875fef0c\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.066641 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities\") pod \"33daa866-c097-4020-9432-0b76875fef0c\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.066745 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kscjj\" (UniqueName: \"kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj\") pod \"33daa866-c097-4020-9432-0b76875fef0c\" (UID: \"33daa866-c097-4020-9432-0b76875fef0c\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.068632 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities" (OuterVolumeSpecName: "utilities") pod "33daa866-c097-4020-9432-0b76875fef0c" (UID: "33daa866-c097-4020-9432-0b76875fef0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.074329 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj" (OuterVolumeSpecName: "kube-api-access-kscjj") pod "33daa866-c097-4020-9432-0b76875fef0c" (UID: "33daa866-c097-4020-9432-0b76875fef0c"). InnerVolumeSpecName "kube-api-access-kscjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.168904 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.168941 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kscjj\" (UniqueName: \"kubernetes.io/projected/33daa866-c097-4020-9432-0b76875fef0c-kube-api-access-kscjj\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.178950 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33daa866-c097-4020-9432-0b76875fef0c" (UID: "33daa866-c097-4020-9432-0b76875fef0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.270745 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33daa866-c097-4020-9432-0b76875fef0c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.335129 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.372677 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnntl\" (UniqueName: \"kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl\") pod \"30b1b73f-cded-42c6-8c1c-104d51d19e15\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.372855 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities\") pod \"30b1b73f-cded-42c6-8c1c-104d51d19e15\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.372955 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content\") pod \"30b1b73f-cded-42c6-8c1c-104d51d19e15\" (UID: \"30b1b73f-cded-42c6-8c1c-104d51d19e15\") " Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.377462 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities" (OuterVolumeSpecName: "utilities") pod "30b1b73f-cded-42c6-8c1c-104d51d19e15" (UID: "30b1b73f-cded-42c6-8c1c-104d51d19e15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.382659 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl" (OuterVolumeSpecName: "kube-api-access-rnntl") pod "30b1b73f-cded-42c6-8c1c-104d51d19e15" (UID: "30b1b73f-cded-42c6-8c1c-104d51d19e15"). InnerVolumeSpecName "kube-api-access-rnntl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.475104 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnntl\" (UniqueName: \"kubernetes.io/projected/30b1b73f-cded-42c6-8c1c-104d51d19e15-kube-api-access-rnntl\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.475150 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.491264 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30b1b73f-cded-42c6-8c1c-104d51d19e15" (UID: "30b1b73f-cded-42c6-8c1c-104d51d19e15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.536800 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.537148 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8n8s" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="registry-server" containerID="cri-o://ce2d671d2deea5e40e642f830f72e4d243ff4b8d0e9a96ee345ab2340929f1a4" gracePeriod=2 Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.576750 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b1b73f-cded-42c6-8c1c-104d51d19e15-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.740156 4761 generic.go:334] "Generic (PLEG): container finished" podID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerID="0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a" exitCode=0 Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.740232 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerDied","Data":"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a"} Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.740262 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwb7t" event={"ID":"30b1b73f-cded-42c6-8c1c-104d51d19e15","Type":"ContainerDied","Data":"8e512a5babf65e1ebc7425bbd57a02e4ce8103e38d4b01181277ac17398b30fe"} Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.740284 4761 scope.go:117] "RemoveContainer" containerID="0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.740430 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwb7t" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.747201 4761 generic.go:334] "Generic (PLEG): container finished" podID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerID="ce2d671d2deea5e40e642f830f72e4d243ff4b8d0e9a96ee345ab2340929f1a4" exitCode=0 Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.747264 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerDied","Data":"ce2d671d2deea5e40e642f830f72e4d243ff4b8d0e9a96ee345ab2340929f1a4"} Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.749687 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vb27h" event={"ID":"33daa866-c097-4020-9432-0b76875fef0c","Type":"ContainerDied","Data":"a6aa6f892a7dca46bfd360026dc3955f7be5a700577aa98b3d3fea10813779fd"} Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.749785 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vb27h" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.793002 4761 scope.go:117] "RemoveContainer" containerID="b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.797873 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.807877 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fwb7t"] Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.816362 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.823530 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vb27h"] Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.823799 4761 scope.go:117] "RemoveContainer" containerID="ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.863604 4761 scope.go:117] "RemoveContainer" containerID="0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a" Nov 25 23:22:19 crc kubenswrapper[4761]: E1125 23:22:19.864012 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a\": container with ID starting with 0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a not found: ID does not exist" containerID="0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.864061 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a"} err="failed to get container status \"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a\": rpc error: code = NotFound desc = could not find container \"0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a\": container with ID starting with 0b15327bfd74fb70883aa53ca34adbbed852a65a5675224e826e65d8ed30be6a not found: ID does not exist" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.864093 4761 scope.go:117] "RemoveContainer" containerID="b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc" Nov 25 23:22:19 crc kubenswrapper[4761]: E1125 23:22:19.864577 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc\": container with ID starting with b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc not found: ID does not exist" containerID="b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.864607 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc"} err="failed to get container status \"b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc\": rpc error: code = NotFound desc = could not find container \"b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc\": container with ID starting with b850c255921a4f03a27a6f4faa701f25c0b25142da66ffc65fc1fae7c16c84dc not found: ID does not exist" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.864630 4761 scope.go:117] "RemoveContainer" containerID="ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48" Nov 25 23:22:19 crc kubenswrapper[4761]: E1125 23:22:19.864998 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48\": container with ID starting with ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48 not found: ID does not exist" containerID="ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.865026 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48"} err="failed to get container status \"ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48\": rpc error: code = NotFound desc = could not find container \"ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48\": container with ID starting with ed401f2cfe87e9ba389281b581a744f759fbc1fec88149ab4d81987aa39e1b48 not found: ID does not exist" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.865045 4761 scope.go:117] "RemoveContainer" containerID="42c97cc799995e28f035584a03b9388ebaf0e07942708477ecae89ddf115bed5" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.903270 4761 scope.go:117] "RemoveContainer" containerID="13fbe95f6c4583ee40abe0c599356ca54b4766716b57757a107f5f06d445990b" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.943690 4761 scope.go:117] "RemoveContainer" containerID="0777d443f53dc2de9bc9b35674299a8ca4ac91c306022dc17d166e468373ba13" Nov 25 23:22:19 crc kubenswrapper[4761]: I1125 23:22:19.996513 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.099155 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6g58\" (UniqueName: \"kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58\") pod \"5ab37abb-3734-45cb-88b3-fa726716d5e0\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.099242 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") pod \"5ab37abb-3734-45cb-88b3-fa726716d5e0\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.099343 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities\") pod \"5ab37abb-3734-45cb-88b3-fa726716d5e0\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.101130 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities" (OuterVolumeSpecName: "utilities") pod "5ab37abb-3734-45cb-88b3-fa726716d5e0" (UID: "5ab37abb-3734-45cb-88b3-fa726716d5e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.105025 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58" (OuterVolumeSpecName: "kube-api-access-b6g58") pod "5ab37abb-3734-45cb-88b3-fa726716d5e0" (UID: "5ab37abb-3734-45cb-88b3-fa726716d5e0"). InnerVolumeSpecName "kube-api-access-b6g58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.148909 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.149258 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4c4l4" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="registry-server" containerID="cri-o://76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6" gracePeriod=2 Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.200505 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ab37abb-3734-45cb-88b3-fa726716d5e0" (UID: "5ab37abb-3734-45cb-88b3-fa726716d5e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.200741 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") pod \"5ab37abb-3734-45cb-88b3-fa726716d5e0\" (UID: \"5ab37abb-3734-45cb-88b3-fa726716d5e0\") " Nov 25 23:22:20 crc kubenswrapper[4761]: W1125 23:22:20.201101 4761 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5ab37abb-3734-45cb-88b3-fa726716d5e0/volumes/kubernetes.io~empty-dir/catalog-content Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.201111 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6g58\" (UniqueName: \"kubernetes.io/projected/5ab37abb-3734-45cb-88b3-fa726716d5e0-kube-api-access-b6g58\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.201116 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ab37abb-3734-45cb-88b3-fa726716d5e0" (UID: "5ab37abb-3734-45cb-88b3-fa726716d5e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.201128 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: E1125 23:22:20.237146 4761 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod032c8467_d8b9_4345_88a7_662e363a3174.slice/crio-conmon-76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6.scope\": RecentStats: unable to find data in memory cache]" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.303058 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab37abb-3734-45cb-88b3-fa726716d5e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.643420 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.709606 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj2bx\" (UniqueName: \"kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx\") pod \"032c8467-d8b9-4345-88a7-662e363a3174\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.709743 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities\") pod \"032c8467-d8b9-4345-88a7-662e363a3174\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.709805 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content\") pod \"032c8467-d8b9-4345-88a7-662e363a3174\" (UID: \"032c8467-d8b9-4345-88a7-662e363a3174\") " Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.710842 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities" (OuterVolumeSpecName: "utilities") pod "032c8467-d8b9-4345-88a7-662e363a3174" (UID: "032c8467-d8b9-4345-88a7-662e363a3174"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.714641 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx" (OuterVolumeSpecName: "kube-api-access-tj2bx") pod "032c8467-d8b9-4345-88a7-662e363a3174" (UID: "032c8467-d8b9-4345-88a7-662e363a3174"). InnerVolumeSpecName "kube-api-access-tj2bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.774801 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8n8s" event={"ID":"5ab37abb-3734-45cb-88b3-fa726716d5e0","Type":"ContainerDied","Data":"96b8dc246b7daffb6720635fd961188f9b73f1ddd5a3e910dfbb1d6e13a3c5a7"} Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.774853 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8n8s" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.774868 4761 scope.go:117] "RemoveContainer" containerID="ce2d671d2deea5e40e642f830f72e4d243ff4b8d0e9a96ee345ab2340929f1a4" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.779770 4761 generic.go:334] "Generic (PLEG): container finished" podID="032c8467-d8b9-4345-88a7-662e363a3174" containerID="76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6" exitCode=0 Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.779841 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerDied","Data":"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6"} Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.779873 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4c4l4" event={"ID":"032c8467-d8b9-4345-88a7-662e363a3174","Type":"ContainerDied","Data":"677de3c35dd665417de67eabc0ebc40d6a67d1d9841242c0647231f92e4ec47a"} Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.779942 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4c4l4" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.809935 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "032c8467-d8b9-4345-88a7-662e363a3174" (UID: "032c8467-d8b9-4345-88a7-662e363a3174"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.811156 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj2bx\" (UniqueName: \"kubernetes.io/projected/032c8467-d8b9-4345-88a7-662e363a3174-kube-api-access-tj2bx\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.811187 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.811200 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032c8467-d8b9-4345-88a7-662e363a3174-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.817963 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.819023 4761 scope.go:117] "RemoveContainer" containerID="3c034ac3e26e672c8994902535c9d35851126ea915d738f47f58f0ddaadd3b14" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.828207 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8n8s"] Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.845785 4761 scope.go:117] "RemoveContainer" containerID="7a424506ad005f294b44d33035f3d4fb2a1c0b568e46ce474abbef45903f2f19" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.873055 4761 scope.go:117] "RemoveContainer" containerID="76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.888010 4761 scope.go:117] "RemoveContainer" containerID="b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.927911 4761 scope.go:117] "RemoveContainer" containerID="fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.965994 4761 scope.go:117] "RemoveContainer" containerID="76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6" Nov 25 23:22:20 crc kubenswrapper[4761]: E1125 23:22:20.968192 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6\": container with ID starting with 76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6 not found: ID does not exist" containerID="76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.968444 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6"} err="failed to get container status \"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6\": rpc error: code = NotFound desc = could not find container \"76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6\": container with ID starting with 76df97d5312bbd771df9dca907b055615f52f3fe00dab93ea154dfa8787deaf6 not found: ID does not exist" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.968475 4761 scope.go:117] "RemoveContainer" containerID="b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859" Nov 25 23:22:20 crc kubenswrapper[4761]: E1125 23:22:20.972169 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859\": container with ID starting with b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859 not found: ID does not exist" containerID="b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.972221 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859"} err="failed to get container status \"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859\": rpc error: code = NotFound desc = could not find container \"b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859\": container with ID starting with b5aadda3dabfd2176eab235f486723f32833b34e72345bdbce283421053fa859 not found: ID does not exist" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.972256 4761 scope.go:117] "RemoveContainer" containerID="fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140" Nov 25 23:22:20 crc kubenswrapper[4761]: E1125 23:22:20.976166 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140\": container with ID starting with fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140 not found: ID does not exist" containerID="fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140" Nov 25 23:22:20 crc kubenswrapper[4761]: I1125 23:22:20.976214 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140"} err="failed to get container status \"fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140\": rpc error: code = NotFound desc = could not find container \"fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140\": container with ID starting with fe837b945e16b9c53b42fd10f76c74c3dd2b38fae19613e31977b3af1ba69140 not found: ID does not exist" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.020040 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" path="/var/lib/kubelet/pods/30b1b73f-cded-42c6-8c1c-104d51d19e15/volumes" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.020777 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33daa866-c097-4020-9432-0b76875fef0c" path="/var/lib/kubelet/pods/33daa866-c097-4020-9432-0b76875fef0c/volumes" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.021433 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" path="/var/lib/kubelet/pods/5ab37abb-3734-45cb-88b3-fa726716d5e0/volumes" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.100874 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.106401 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4c4l4"] Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.127807 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.127902 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.127974 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.128913 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.129312 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" gracePeriod=600 Nov 25 23:22:21 crc kubenswrapper[4761]: E1125 23:22:21.254124 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.798080 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" exitCode=0 Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.798175 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251"} Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.798257 4761 scope.go:117] "RemoveContainer" containerID="45d380506524b52ea4ad1ad414e953d43b055e8a693d37ed77a271f4bbee6115" Nov 25 23:22:21 crc kubenswrapper[4761]: I1125 23:22:21.799054 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:22:21 crc kubenswrapper[4761]: E1125 23:22:21.799439 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:22:23 crc kubenswrapper[4761]: I1125 23:22:23.033745 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032c8467-d8b9-4345-88a7-662e363a3174" path="/var/lib/kubelet/pods/032c8467-d8b9-4345-88a7-662e363a3174/volumes" Nov 25 23:22:32 crc kubenswrapper[4761]: I1125 23:22:32.011857 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:22:32 crc kubenswrapper[4761]: E1125 23:22:32.012951 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:22:47 crc kubenswrapper[4761]: I1125 23:22:47.011166 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:22:47 crc kubenswrapper[4761]: E1125 23:22:47.012167 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:22:47 crc kubenswrapper[4761]: I1125 23:22:47.445852 4761 scope.go:117] "RemoveContainer" containerID="774285aecbee5bfcad39e7ed1cbfb99bbb4ac2059dfae882e8e6ee368d0d4481" Nov 25 23:22:47 crc kubenswrapper[4761]: I1125 23:22:47.514169 4761 scope.go:117] "RemoveContainer" containerID="f038d6766a854ff30e58dd0bccbecc637071ad169e27825d17deb52063b02949" Nov 25 23:22:58 crc kubenswrapper[4761]: I1125 23:22:58.012577 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:22:58 crc kubenswrapper[4761]: E1125 23:22:58.013946 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:23:10 crc kubenswrapper[4761]: I1125 23:23:10.011623 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:23:10 crc kubenswrapper[4761]: E1125 23:23:10.012828 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:23:21 crc kubenswrapper[4761]: I1125 23:23:21.011946 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:23:21 crc kubenswrapper[4761]: E1125 23:23:21.013146 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:23:33 crc kubenswrapper[4761]: I1125 23:23:33.011991 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:23:33 crc kubenswrapper[4761]: E1125 23:23:33.013302 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:23:48 crc kubenswrapper[4761]: I1125 23:23:48.011211 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:23:48 crc kubenswrapper[4761]: E1125 23:23:48.012514 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:24:03 crc kubenswrapper[4761]: I1125 23:24:03.011828 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:24:03 crc kubenswrapper[4761]: E1125 23:24:03.013146 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:24:18 crc kubenswrapper[4761]: I1125 23:24:18.012229 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:24:18 crc kubenswrapper[4761]: E1125 23:24:18.013191 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.153818 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155210 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155248 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155276 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155294 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155321 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155339 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155372 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155391 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155413 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155430 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155468 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155486 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155510 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155526 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155558 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155573 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155599 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155612 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155630 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155643 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155667 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155680 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155697 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155757 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155793 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155809 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155833 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155846 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155862 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155875 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155901 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155914 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155931 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155944 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155966 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.155979 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="extract-content" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.155996 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156008 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="extract-utilities" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.156025 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156040 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: E1125 23:24:24.156053 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156065 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156316 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="933d00e9-8ab4-4a1b-a7d1-ae18bf7e2a76" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156351 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="33daa866-c097-4020-9432-0b76875fef0c" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156377 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="346df741-cf07-4002-a8b6-5069c015aee3" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156395 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc9672bd-1e5e-4ef7-8713-145c1a8a6cd6" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156412 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="032c8467-d8b9-4345-88a7-662e363a3174" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156434 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b1b73f-cded-42c6-8c1c-104d51d19e15" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.156457 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab37abb-3734-45cb-88b3-fa726716d5e0" containerName="registry-server" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.158284 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.169421 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.195769 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.197107 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.201243 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.201388 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-lmqzr" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.201507 4761 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.202578 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.204467 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.208355 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.208539 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7cb\" (UniqueName: \"kubernetes.io/projected/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-kube-api-access-md7cb\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.208573 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-scripts\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.208617 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config-secret\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.309739 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwz46\" (UniqueName: \"kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.309872 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.309932 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7cb\" (UniqueName: \"kubernetes.io/projected/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-kube-api-access-md7cb\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.309959 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-scripts\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.309996 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310022 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310046 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310087 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config-secret\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310110 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310133 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310177 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310216 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310336 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310436 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310533 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310560 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310597 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.310627 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.311057 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.312213 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-scripts\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.315728 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-openstack-config-secret\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.331900 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7cb\" (UniqueName: \"kubernetes.io/projected/77988aba-9a8b-4f63-b35a-6ecb7a4c8309-kube-api-access-md7cb\") pod \"openstackclient\" (UID: \"77988aba-9a8b-4f63-b35a-6ecb7a4c8309\") " pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412178 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412227 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412253 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412275 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412328 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwz46\" (UniqueName: \"kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412358 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412382 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412401 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412418 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412442 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412456 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412477 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412491 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412498 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412587 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412602 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412502 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412646 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412670 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412655 4761 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412713 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412853 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.412933 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.413356 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.413417 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.416244 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.416367 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.431613 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwz46\" (UniqueName: \"kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.438986 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.440585 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.482938 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.519618 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.730458 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 23:24:24 crc kubenswrapper[4761]: W1125 23:24:24.745666 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77988aba_9a8b_4f63_b35a_6ecb7a4c8309.slice/crio-125be43afb795b1e9fc04754ff51b5810b35ceec82e61e479b5b2cac8bb39264 WatchSource:0}: Error finding container 125be43afb795b1e9fc04754ff51b5810b35ceec82e61e479b5b2cac8bb39264: Status 404 returned error can't find the container with id 125be43afb795b1e9fc04754ff51b5810b35ceec82e61e479b5b2cac8bb39264 Nov 25 23:24:24 crc kubenswrapper[4761]: I1125 23:24:24.809471 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:24:24 crc kubenswrapper[4761]: W1125 23:24:24.811851 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8278d408_1b3e_478a_9d08_0e964b0b0ffc.slice/crio-89af942a32150ad2d8b90e8e2c1ac7939a29a483f0df5b5b69c49d13e21c6809 WatchSource:0}: Error finding container 89af942a32150ad2d8b90e8e2c1ac7939a29a483f0df5b5b69c49d13e21c6809: Status 404 returned error can't find the container with id 89af942a32150ad2d8b90e8e2c1ac7939a29a483f0df5b5b69c49d13e21c6809 Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.092083 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerStarted","Data":"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183"} Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.092233 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerStarted","Data":"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef"} Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.092244 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerStarted","Data":"89af942a32150ad2d8b90e8e2c1ac7939a29a483f0df5b5b69c49d13e21c6809"} Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.093819 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"77988aba-9a8b-4f63-b35a-6ecb7a4c8309","Type":"ContainerStarted","Data":"c565a3f42876cee638c1eb3109abba446ae342265097293385c144ccdf87e98b"} Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.093868 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"77988aba-9a8b-4f63-b35a-6ecb7a4c8309","Type":"ContainerStarted","Data":"125be43afb795b1e9fc04754ff51b5810b35ceec82e61e479b5b2cac8bb39264"} Nov 25 23:24:25 crc kubenswrapper[4761]: I1125 23:24:25.120517 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.120499086 podStartE2EDuration="1.120499086s" podCreationTimestamp="2025-11-25 23:24:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:24:25.109416859 +0000 UTC m=+1660.842762714" watchObservedRunningTime="2025-11-25 23:24:25.120499086 +0000 UTC m=+1660.853844921" Nov 25 23:24:26 crc kubenswrapper[4761]: I1125 23:24:26.106274 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerStarted","Data":"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01"} Nov 25 23:24:26 crc kubenswrapper[4761]: I1125 23:24:26.157030 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.156993994 podStartE2EDuration="2.156993994s" podCreationTimestamp="2025-11-25 23:24:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 23:24:26.14787046 +0000 UTC m=+1661.881216375" watchObservedRunningTime="2025-11-25 23:24:26.156993994 +0000 UTC m=+1661.890339879" Nov 25 23:24:29 crc kubenswrapper[4761]: I1125 23:24:29.012059 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:24:29 crc kubenswrapper[4761]: E1125 23:24:29.013120 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.483810 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.484950 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.484977 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.515787 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.529592 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:34 crc kubenswrapper[4761]: I1125 23:24:34.553993 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.197547 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.197673 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.197724 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.218564 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.219172 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:35 crc kubenswrapper[4761]: I1125 23:24:35.221078 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:24:41 crc kubenswrapper[4761]: I1125 23:24:41.011958 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:24:41 crc kubenswrapper[4761]: E1125 23:24:41.013162 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:24:53 crc kubenswrapper[4761]: I1125 23:24:53.012018 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:24:53 crc kubenswrapper[4761]: E1125 23:24:53.013051 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:25:08 crc kubenswrapper[4761]: I1125 23:25:08.011115 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:25:08 crc kubenswrapper[4761]: E1125 23:25:08.012016 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:25:20 crc kubenswrapper[4761]: I1125 23:25:20.011396 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:25:20 crc kubenswrapper[4761]: E1125 23:25:20.013565 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:25:31 crc kubenswrapper[4761]: I1125 23:25:31.011317 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:25:31 crc kubenswrapper[4761]: E1125 23:25:31.012077 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:25:31 crc kubenswrapper[4761]: I1125 23:25:31.899055 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:25:31 crc kubenswrapper[4761]: I1125 23:25:31.899881 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-log" containerID="cri-o://2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" gracePeriod=30 Nov 25 23:25:31 crc kubenswrapper[4761]: I1125 23:25:31.900040 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-api" containerID="cri-o://5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" gracePeriod=30 Nov 25 23:25:31 crc kubenswrapper[4761]: I1125 23:25:31.900014 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-httpd" containerID="cri-o://019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" gracePeriod=30 Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.811133 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971376 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971441 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971471 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971490 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971581 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971613 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971693 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971756 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971793 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwz46\" (UniqueName: \"kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971860 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971886 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs" (OuterVolumeSpecName: "logs") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971905 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971946 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.971987 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972026 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972090 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972126 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys" (OuterVolumeSpecName: "sys") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972140 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules\") pod \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\" (UID: \"8278d408-1b3e-478a-9d08-0e964b0b0ffc\") " Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972291 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972295 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972395 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run" (OuterVolumeSpecName: "run") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972439 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev" (OuterVolumeSpecName: "dev") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972473 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972839 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972862 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972879 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972895 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972913 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972928 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972944 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972960 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8278d408-1b3e-478a-9d08-0e964b0b0ffc-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.972977 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8278d408-1b3e-478a-9d08-0e964b0b0ffc-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975486 4761 generic.go:334] "Generic (PLEG): container finished" podID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" exitCode=0 Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975530 4761 generic.go:334] "Generic (PLEG): container finished" podID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" exitCode=0 Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975546 4761 generic.go:334] "Generic (PLEG): container finished" podID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" exitCode=143 Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975563 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975574 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerDied","Data":"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01"} Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975615 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerDied","Data":"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183"} Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975634 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerDied","Data":"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef"} Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975653 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"8278d408-1b3e-478a-9d08-0e964b0b0ffc","Type":"ContainerDied","Data":"89af942a32150ad2d8b90e8e2c1ac7939a29a483f0df5b5b69c49d13e21c6809"} Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.975680 4761 scope.go:117] "RemoveContainer" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.978278 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.980124 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.980713 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts" (OuterVolumeSpecName: "scripts") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:32 crc kubenswrapper[4761]: I1125 23:25:32.981946 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46" (OuterVolumeSpecName: "kube-api-access-hwz46") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "kube-api-access-hwz46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.012852 4761 scope.go:117] "RemoveContainer" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.074871 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.074928 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.074951 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwz46\" (UniqueName: \"kubernetes.io/projected/8278d408-1b3e-478a-9d08-0e964b0b0ffc-kube-api-access-hwz46\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.074972 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.075764 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data" (OuterVolumeSpecName: "config-data") pod "8278d408-1b3e-478a-9d08-0e964b0b0ffc" (UID: "8278d408-1b3e-478a-9d08-0e964b0b0ffc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.091180 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.091514 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.151035 4761 scope.go:117] "RemoveContainer" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.172954 4761 scope.go:117] "RemoveContainer" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.173431 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": container with ID starting with 5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01 not found: ID does not exist" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.173509 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01"} err="failed to get container status \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": rpc error: code = NotFound desc = could not find container \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": container with ID starting with 5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.173550 4761 scope.go:117] "RemoveContainer" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.174078 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": container with ID starting with 019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183 not found: ID does not exist" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174110 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183"} err="failed to get container status \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": rpc error: code = NotFound desc = could not find container \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": container with ID starting with 019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174132 4761 scope.go:117] "RemoveContainer" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.174429 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": container with ID starting with 2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef not found: ID does not exist" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174476 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef"} err="failed to get container status \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": rpc error: code = NotFound desc = could not find container \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": container with ID starting with 2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174506 4761 scope.go:117] "RemoveContainer" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174852 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01"} err="failed to get container status \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": rpc error: code = NotFound desc = could not find container \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": container with ID starting with 5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.174876 4761 scope.go:117] "RemoveContainer" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.175252 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183"} err="failed to get container status \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": rpc error: code = NotFound desc = could not find container \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": container with ID starting with 019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.175327 4761 scope.go:117] "RemoveContainer" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.175674 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef"} err="failed to get container status \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": rpc error: code = NotFound desc = could not find container \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": container with ID starting with 2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.175717 4761 scope.go:117] "RemoveContainer" containerID="5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176071 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01"} err="failed to get container status \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": rpc error: code = NotFound desc = could not find container \"5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01\": container with ID starting with 5028299cb10b00c25fe49004b7d4eefa478545851552ae80bbaa05b83c029a01 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176110 4761 scope.go:117] "RemoveContainer" containerID="019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176298 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8278d408-1b3e-478a-9d08-0e964b0b0ffc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176340 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176360 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176422 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183"} err="failed to get container status \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": rpc error: code = NotFound desc = could not find container \"019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183\": container with ID starting with 019be69652026071f8f4d02ec2e67dbc7137d01d85ac973ae75dc036a701a183 not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176450 4761 scope.go:117] "RemoveContainer" containerID="2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.176805 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef"} err="failed to get container status \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": rpc error: code = NotFound desc = could not find container \"2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef\": container with ID starting with 2b7e69f7a207a085876539b779655e16f49729ee48154b454f88a5f81b9964ef not found: ID does not exist" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.242188 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-l2dbz"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.248964 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-l2dbz"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.285812 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance32e9-account-delete-jjx2t"] Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.286200 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-log" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286219 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-log" Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.286239 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-httpd" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286247 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-httpd" Nov 25 23:25:33 crc kubenswrapper[4761]: E1125 23:25:33.286258 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-api" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286265 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-api" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286745 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-log" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286832 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-api" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.286856 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" containerName="glance-httpd" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.287674 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.298812 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance32e9-account-delete-jjx2t"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.328656 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.337045 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.394272 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.394518 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-log" containerID="cri-o://9c7761d642885dd8874a392f404719008ac20fa5e8e4a6cd42e432c165875c45" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.394622 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-api" containerID="cri-o://0cf9efa34066c0d5e1299152b15e28d04ca1fd68c8c07f9d0e3b179e5da98363" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.394647 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-httpd" containerID="cri-o://d1debbdaa27b4f5b6e8586b21ec3bac136b5dde89aa9a793489a29937c1cf78b" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.459485 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.459805 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-log" containerID="cri-o://3131aee74aa040be6a00b9572c38e93263950b5f35b9f14361d2e59fde5da6a0" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.459885 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-api" containerID="cri-o://5910c6a426abd2c113d99b56a40667340d1976fbe384d403a6ec96fedbdfcbb3" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.459924 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-httpd" containerID="cri-o://302494bd6d6e85ec65176c3fbc7331edaa2c254340853f2248669e146ca33d9b" gracePeriod=30 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.480087 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.480139 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5dzh\" (UniqueName: \"kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.580971 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5dzh\" (UniqueName: \"kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.581184 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.581938 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.611728 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5dzh\" (UniqueName: \"kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh\") pod \"glance32e9-account-delete-jjx2t\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.613656 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984558 4761 generic.go:334] "Generic (PLEG): container finished" podID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerID="0cf9efa34066c0d5e1299152b15e28d04ca1fd68c8c07f9d0e3b179e5da98363" exitCode=0 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984881 4761 generic.go:334] "Generic (PLEG): container finished" podID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerID="d1debbdaa27b4f5b6e8586b21ec3bac136b5dde89aa9a793489a29937c1cf78b" exitCode=0 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984893 4761 generic.go:334] "Generic (PLEG): container finished" podID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerID="9c7761d642885dd8874a392f404719008ac20fa5e8e4a6cd42e432c165875c45" exitCode=143 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984657 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerDied","Data":"0cf9efa34066c0d5e1299152b15e28d04ca1fd68c8c07f9d0e3b179e5da98363"} Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984933 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerDied","Data":"d1debbdaa27b4f5b6e8586b21ec3bac136b5dde89aa9a793489a29937c1cf78b"} Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.984949 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerDied","Data":"9c7761d642885dd8874a392f404719008ac20fa5e8e4a6cd42e432c165875c45"} Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997236 4761 generic.go:334] "Generic (PLEG): container finished" podID="498370d5-ca44-4d35-89b2-9870dff0374a" containerID="5910c6a426abd2c113d99b56a40667340d1976fbe384d403a6ec96fedbdfcbb3" exitCode=0 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997261 4761 generic.go:334] "Generic (PLEG): container finished" podID="498370d5-ca44-4d35-89b2-9870dff0374a" containerID="302494bd6d6e85ec65176c3fbc7331edaa2c254340853f2248669e146ca33d9b" exitCode=0 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997268 4761 generic.go:334] "Generic (PLEG): container finished" podID="498370d5-ca44-4d35-89b2-9870dff0374a" containerID="3131aee74aa040be6a00b9572c38e93263950b5f35b9f14361d2e59fde5da6a0" exitCode=143 Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997305 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerDied","Data":"5910c6a426abd2c113d99b56a40667340d1976fbe384d403a6ec96fedbdfcbb3"} Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997330 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerDied","Data":"302494bd6d6e85ec65176c3fbc7331edaa2c254340853f2248669e146ca33d9b"} Nov 25 23:25:33 crc kubenswrapper[4761]: I1125 23:25:33.997338 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerDied","Data":"3131aee74aa040be6a00b9572c38e93263950b5f35b9f14361d2e59fde5da6a0"} Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.039142 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance32e9-account-delete-jjx2t"] Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.263668 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.267654 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395806 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395875 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jcvv\" (UniqueName: \"kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395914 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395932 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395957 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.395998 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396014 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396035 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396052 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396070 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396090 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396111 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396136 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396151 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396185 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396200 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nrsq\" (UniqueName: \"kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396218 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396234 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396250 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396262 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396278 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396297 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396320 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396335 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396350 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396387 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396399 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev\") pod \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\" (UID: \"f66b46b7-0b13-4ae9-a945-ee8d8f424492\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396420 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme\") pod \"498370d5-ca44-4d35-89b2-9870dff0374a\" (UID: \"498370d5-ca44-4d35-89b2-9870dff0374a\") " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396680 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.396792 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.398331 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.398922 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399069 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run" (OuterVolumeSpecName: "run") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399202 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev" (OuterVolumeSpecName: "dev") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399271 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys" (OuterVolumeSpecName: "sys") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399631 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev" (OuterVolumeSpecName: "dev") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399647 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399682 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399850 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399885 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys" (OuterVolumeSpecName: "sys") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399895 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs" (OuterVolumeSpecName: "logs") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399907 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399930 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399951 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run" (OuterVolumeSpecName: "run") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.399971 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.400072 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs" (OuterVolumeSpecName: "logs") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.402312 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts" (OuterVolumeSpecName: "scripts") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.402594 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts" (OuterVolumeSpecName: "scripts") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.402905 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv" (OuterVolumeSpecName: "kube-api-access-4jcvv") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "kube-api-access-4jcvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.403684 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq" (OuterVolumeSpecName: "kube-api-access-8nrsq") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "kube-api-access-8nrsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.403973 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.404284 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.404337 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.407580 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.465017 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data" (OuterVolumeSpecName: "config-data") pod "f66b46b7-0b13-4ae9-a945-ee8d8f424492" (UID: "f66b46b7-0b13-4ae9-a945-ee8d8f424492"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.468538 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data" (OuterVolumeSpecName: "config-data") pod "498370d5-ca44-4d35-89b2-9870dff0374a" (UID: "498370d5-ca44-4d35-89b2-9870dff0374a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.498737 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.498921 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499081 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499210 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499333 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jcvv\" (UniqueName: \"kubernetes.io/projected/f66b46b7-0b13-4ae9-a945-ee8d8f424492-kube-api-access-4jcvv\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499462 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499574 4761 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499687 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499850 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.499995 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500113 4761 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-sys\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500222 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500331 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500438 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500591 4761 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500773 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.500919 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.501936 4761 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498370d5-ca44-4d35-89b2-9870dff0374a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.501979 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nrsq\" (UniqueName: \"kubernetes.io/projected/498370d5-ca44-4d35-89b2-9870dff0374a-kube-api-access-8nrsq\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.501996 4761 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498370d5-ca44-4d35-89b2-9870dff0374a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502010 4761 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f66b46b7-0b13-4ae9-a945-ee8d8f424492-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502043 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502056 4761 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66b46b7-0b13-4ae9-a945-ee8d8f424492-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502068 4761 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f66b46b7-0b13-4ae9-a945-ee8d8f424492-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502085 4761 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502097 4761 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-dev\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502109 4761 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.502121 4761 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/498370d5-ca44-4d35-89b2-9870dff0374a-run\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.519354 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.525826 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.531949 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.537812 4761 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.603927 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.604149 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.604343 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:34 crc kubenswrapper[4761]: I1125 23:25:34.604479 4761 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.017867 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.024851 4761 generic.go:334] "Generic (PLEG): container finished" podID="f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" containerID="6f36081412228b01edd95f9e1fd24113d067f09f15a3335e5b91037a70fa9254" exitCode=0 Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.027909 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06655c0a-237e-4a10-b264-55c63afb1198" path="/var/lib/kubelet/pods/06655c0a-237e-4a10-b264-55c63afb1198/volumes" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.028738 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8278d408-1b3e-478a-9d08-0e964b0b0ffc" path="/var/lib/kubelet/pods/8278d408-1b3e-478a-9d08-0e964b0b0ffc/volumes" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.029580 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f66b46b7-0b13-4ae9-a945-ee8d8f424492","Type":"ContainerDied","Data":"c32da9efae5ec1ab37ef0ee926acdf1a78d103c49f689b160b3fe8622bd57901"} Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.029614 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" event={"ID":"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc","Type":"ContainerDied","Data":"6f36081412228b01edd95f9e1fd24113d067f09f15a3335e5b91037a70fa9254"} Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.029630 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" event={"ID":"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc","Type":"ContainerStarted","Data":"2f97e20e5476e24a5491406af4bad0a939d5c9179c53ad9cb2235f71298e8f19"} Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.029649 4761 scope.go:117] "RemoveContainer" containerID="0cf9efa34066c0d5e1299152b15e28d04ca1fd68c8c07f9d0e3b179e5da98363" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.031403 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.031451 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"498370d5-ca44-4d35-89b2-9870dff0374a","Type":"ContainerDied","Data":"038fd72f5bea03661fa2050cda00e684f280c6fdf0b1bb616c4d6d1bf70c63b7"} Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.100449 4761 scope.go:117] "RemoveContainer" containerID="d1debbdaa27b4f5b6e8586b21ec3bac136b5dde89aa9a793489a29937c1cf78b" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.114439 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.127013 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.132677 4761 scope.go:117] "RemoveContainer" containerID="9c7761d642885dd8874a392f404719008ac20fa5e8e4a6cd42e432c165875c45" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.134660 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.141925 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.152439 4761 scope.go:117] "RemoveContainer" containerID="5910c6a426abd2c113d99b56a40667340d1976fbe384d403a6ec96fedbdfcbb3" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.172439 4761 scope.go:117] "RemoveContainer" containerID="302494bd6d6e85ec65176c3fbc7331edaa2c254340853f2248669e146ca33d9b" Nov 25 23:25:35 crc kubenswrapper[4761]: I1125 23:25:35.193687 4761 scope.go:117] "RemoveContainer" containerID="3131aee74aa040be6a00b9572c38e93263950b5f35b9f14361d2e59fde5da6a0" Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.406798 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.535650 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts\") pod \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.535795 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5dzh\" (UniqueName: \"kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh\") pod \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\" (UID: \"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc\") " Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.536864 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" (UID: "f9835d4c-ffab-45fa-8a52-a47ea24a1cbc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.550615 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh" (OuterVolumeSpecName: "kube-api-access-r5dzh") pod "f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" (UID: "f9835d4c-ffab-45fa-8a52-a47ea24a1cbc"). InnerVolumeSpecName "kube-api-access-r5dzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.651806 4761 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:36 crc kubenswrapper[4761]: I1125 23:25:36.651873 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5dzh\" (UniqueName: \"kubernetes.io/projected/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc-kube-api-access-r5dzh\") on node \"crc\" DevicePath \"\"" Nov 25 23:25:37 crc kubenswrapper[4761]: I1125 23:25:37.025869 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" path="/var/lib/kubelet/pods/498370d5-ca44-4d35-89b2-9870dff0374a/volumes" Nov 25 23:25:37 crc kubenswrapper[4761]: I1125 23:25:37.027262 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" path="/var/lib/kubelet/pods/f66b46b7-0b13-4ae9-a945-ee8d8f424492/volumes" Nov 25 23:25:37 crc kubenswrapper[4761]: I1125 23:25:37.061017 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" event={"ID":"f9835d4c-ffab-45fa-8a52-a47ea24a1cbc","Type":"ContainerDied","Data":"2f97e20e5476e24a5491406af4bad0a939d5c9179c53ad9cb2235f71298e8f19"} Nov 25 23:25:37 crc kubenswrapper[4761]: I1125 23:25:37.061090 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance32e9-account-delete-jjx2t" Nov 25 23:25:37 crc kubenswrapper[4761]: I1125 23:25:37.061119 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f97e20e5476e24a5491406af4bad0a939d5c9179c53ad9cb2235f71298e8f19" Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.301460 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-j2kwl"] Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.313775 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-j2kwl"] Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.334723 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-32e9-account-create-update-f8gtl"] Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.343610 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance32e9-account-delete-jjx2t"] Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.350814 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-32e9-account-create-update-f8gtl"] Nov 25 23:25:38 crc kubenswrapper[4761]: I1125 23:25:38.358890 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance32e9-account-delete-jjx2t"] Nov 25 23:25:39 crc kubenswrapper[4761]: I1125 23:25:39.026419 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01193382-e383-4185-bbd6-af046b634069" path="/var/lib/kubelet/pods/01193382-e383-4185-bbd6-af046b634069/volumes" Nov 25 23:25:39 crc kubenswrapper[4761]: I1125 23:25:39.027795 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f11741e-da8f-4bb7-abf3-038efc7ee2a6" path="/var/lib/kubelet/pods/4f11741e-da8f-4bb7-abf3-038efc7ee2a6/volumes" Nov 25 23:25:39 crc kubenswrapper[4761]: I1125 23:25:39.028925 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" path="/var/lib/kubelet/pods/f9835d4c-ffab-45fa-8a52-a47ea24a1cbc/volumes" Nov 25 23:25:45 crc kubenswrapper[4761]: I1125 23:25:45.014864 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:25:45 crc kubenswrapper[4761]: E1125 23:25:45.015522 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:25:47 crc kubenswrapper[4761]: I1125 23:25:47.704940 4761 scope.go:117] "RemoveContainer" containerID="28aa62ad597020bbf52bcb8be8d3e3c0ad243c54f981e1d57babddda42fc93f0" Nov 25 23:25:47 crc kubenswrapper[4761]: I1125 23:25:47.766274 4761 scope.go:117] "RemoveContainer" containerID="998f4d4c23f0f09f8662f1698cd2759310fd9c740033f4cd870ffb534d76e1f9" Nov 25 23:25:47 crc kubenswrapper[4761]: I1125 23:25:47.802735 4761 scope.go:117] "RemoveContainer" containerID="2dee78dea989d0c0045e9acf721be56a08f8dc0cb5b3273457d25476cc34a728" Nov 25 23:25:58 crc kubenswrapper[4761]: I1125 23:25:58.012208 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:25:58 crc kubenswrapper[4761]: E1125 23:25:58.014233 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.011923 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.012934 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.157045 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-99l4c/must-gather-jkfkv"] Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.157959 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" containerName="mariadb-account-delete" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.157981 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" containerName="mariadb-account-delete" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158004 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158014 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158047 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158056 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158068 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158075 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158084 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158093 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158112 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158119 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: E1125 23:26:13.158131 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158139 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158281 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158295 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158304 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158314 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-log" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158323 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66b46b7-0b13-4ae9-a945-ee8d8f424492" containerName="glance-api" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158337 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9835d4c-ffab-45fa-8a52-a47ea24a1cbc" containerName="mariadb-account-delete" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.158351 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="498370d5-ca44-4d35-89b2-9870dff0374a" containerName="glance-httpd" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.159261 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.161503 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-99l4c"/"kube-root-ca.crt" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.161760 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-99l4c"/"default-dockercfg-7w7wf" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.161932 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-99l4c"/"openshift-service-ca.crt" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.176135 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-99l4c/must-gather-jkfkv"] Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.254732 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgftx\" (UniqueName: \"kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.254795 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.357038 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgftx\" (UniqueName: \"kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.357171 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.357996 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.377801 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgftx\" (UniqueName: \"kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx\") pod \"must-gather-jkfkv\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:13 crc kubenswrapper[4761]: I1125 23:26:13.475258 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:26:14 crc kubenswrapper[4761]: I1125 23:26:14.060631 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-99l4c/must-gather-jkfkv"] Nov 25 23:26:14 crc kubenswrapper[4761]: I1125 23:26:14.072436 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 23:26:14 crc kubenswrapper[4761]: I1125 23:26:14.426806 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-99l4c/must-gather-jkfkv" event={"ID":"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5","Type":"ContainerStarted","Data":"b92435c34ecf67181e6d93b5dfc7675a31eeaabecfd29df7afa6f2e90c8b8242"} Nov 25 23:26:20 crc kubenswrapper[4761]: I1125 23:26:20.229173 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-99l4c/must-gather-jkfkv" event={"ID":"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5","Type":"ContainerStarted","Data":"cf2fb587f81dca73cdc6ceca761d30460da5c708ca2c8cf2a19e3a229813efce"} Nov 25 23:26:20 crc kubenswrapper[4761]: I1125 23:26:20.229689 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-99l4c/must-gather-jkfkv" event={"ID":"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5","Type":"ContainerStarted","Data":"18667b8824eca762b66da1559cc361bee492b8665ff5de80dcdb7c86fd8204c9"} Nov 25 23:26:20 crc kubenswrapper[4761]: I1125 23:26:20.255448 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-99l4c/must-gather-jkfkv" podStartSLOduration=3.110504867 podStartE2EDuration="7.255426618s" podCreationTimestamp="2025-11-25 23:26:13 +0000 UTC" firstStartedPulling="2025-11-25 23:26:14.072219104 +0000 UTC m=+1769.805564939" lastFinishedPulling="2025-11-25 23:26:18.217140855 +0000 UTC m=+1773.950486690" observedRunningTime="2025-11-25 23:26:20.251187923 +0000 UTC m=+1775.984533778" watchObservedRunningTime="2025-11-25 23:26:20.255426618 +0000 UTC m=+1775.988772483" Nov 25 23:26:27 crc kubenswrapper[4761]: I1125 23:26:27.012060 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:26:27 crc kubenswrapper[4761]: E1125 23:26:27.013325 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:26:41 crc kubenswrapper[4761]: I1125 23:26:41.012145 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:26:41 crc kubenswrapper[4761]: E1125 23:26:41.013423 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:26:47 crc kubenswrapper[4761]: I1125 23:26:47.947745 4761 scope.go:117] "RemoveContainer" containerID="500c5ae5a21baefe2028e90c608b4187f9e3bed40d1b7bac534a7eda3bc0b93f" Nov 25 23:26:47 crc kubenswrapper[4761]: I1125 23:26:47.984257 4761 scope.go:117] "RemoveContainer" containerID="2de507af97056a10556e8e772f957e4454dd1625c3a3dd19f2c4398f0d25ca58" Nov 25 23:26:48 crc kubenswrapper[4761]: I1125 23:26:48.001837 4761 scope.go:117] "RemoveContainer" containerID="960bb5300b34104f250886cdbc3862a2dd991ea88b658211150c1e5819545b24" Nov 25 23:26:55 crc kubenswrapper[4761]: I1125 23:26:55.015597 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:26:55 crc kubenswrapper[4761]: E1125 23:26:55.016418 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.035467 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.211884 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/pull/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.240586 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.245358 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/pull/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.400676 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.403323 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/extract/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.421769 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06f09c2701dd20f727237a99bdf246b7079442b8d8d83964701a88d4206hbwq_52cc79c4-6ed7-463c-afae-9e4e9740ddbb/pull/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.596868 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.727242 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/pull/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.752224 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.764981 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/pull/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.876140 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/util/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.912435 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/extract/0.log" Nov 25 23:26:58 crc kubenswrapper[4761]: I1125 23:26:58.915182 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fcxmts_15aaf702-bd98-4dbd-b5e7-296a7546cd5a/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.033735 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/util/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.295403 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.312718 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/util/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.314174 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.490061 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/util/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.502621 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/extract/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.532440 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dpt8vw_205db7be-4e2b-4785-85f4-b5a76f277139/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.644298 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/util/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.820615 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/util/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.844867 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.848265 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/pull/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.993016 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/extract/0.log" Nov 25 23:26:59 crc kubenswrapper[4761]: I1125 23:26:59.994011 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.059604 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bzmkc4_fe7636ab-8296-4e62-bba7-61ccb3507e9f/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.137815 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.342590 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.343491 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.349125 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.462348 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.499797 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/extract/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.534587 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dpfkw_b3986c58-f4d5-430c-a46b-c43b788e0320/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.574725 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.692416 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.722092 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.757737 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.920555 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/util/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.966001 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/extract/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.970878 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cz7nl2_cf932b8b-53ef-4e97-ac65-ec51f2177155/pull/0.log" Nov 25 23:27:00 crc kubenswrapper[4761]: I1125 23:27:00.993357 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/util/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.160326 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/util/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.169086 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/pull/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.206846 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/pull/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.386517 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/pull/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.389976 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/extract/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.398524 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3mqs2q_1206534d-9218-4636-b769-46aab59ed934/util/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.429377 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-66b769fbc8-xf6tp_3daf70a9-4dab-4ace-82c0-9c7e48512cf6/manager/3.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.571947 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-g5dpb_9b3bfb75-fcbe-4b6e-bb0e-8a09630901df/registry-server/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.576563 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-66b769fbc8-xf6tp_3daf70a9-4dab-4ace-82c0-9c7e48512cf6/manager/2.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.638159 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f54679d79-6ldpn_bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b/manager/3.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.724579 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f54679d79-6ldpn_bef48210-b7a0-47d5-b5e2-b5c1d3f8c15b/manager/2.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.771010 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-pkh4w_ba0b63b3-4cf1-42f4-8daf-1b10aca0e4ad/registry-server/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.850676 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c5dd6c96-z8jr7_2e34c701-6bbc-4000-a90a-aeca2333b69b/kube-rbac-proxy/0.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.903026 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c5dd6c96-z8jr7_2e34c701-6bbc-4000-a90a-aeca2333b69b/manager/3.log" Nov 25 23:27:01 crc kubenswrapper[4761]: I1125 23:27:01.951866 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c5dd6c96-z8jr7_2e34c701-6bbc-4000-a90a-aeca2333b69b/manager/2.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.039625 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-4qknl_7fdf0c47-5b58-4cac-8fe2-df4f2e58d797/registry-server/0.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.078880 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-8559b887bb-frnqv_47a4708b-be53-4783-b9c1-9536d90b9e4c/manager/3.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.120589 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-8559b887bb-frnqv_47a4708b-be53-4783-b9c1-9536d90b9e4c/manager/2.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.223988 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-sw22g_464d8029-30b2-4ac1-8820-93dd0f144047/registry-server/0.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.246379 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6bbbcc9f68-dcp42_0717f7d1-d5b3-408e-878e-10c62251448a/manager/3.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.277463 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6bbbcc9f68-dcp42_0717f7d1-d5b3-408e-878e-10c62251448a/manager/2.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.398181 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-tq4fc_a74e9524-d069-42d8-9e98-ec3eed11ec22/registry-server/0.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.420255 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-cskvg_91b5b48e-a642-45e2-a84a-d449123e57bb/operator/3.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.446505 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-cskvg_91b5b48e-a642-45e2-a84a-d449123e57bb/operator/2.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.580523 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-srnc9_43a81c5c-7b19-414c-9e76-9b63aaa0dc14/registry-server/0.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.584606 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6c6d6c68d5-v5cfc_8aba4458-8023-4b9f-8504-6fab49f5250c/manager/3.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.614257 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6c6d6c68d5-v5cfc_8aba4458-8023-4b9f-8504-6fab49f5250c/manager/2.log" Nov 25 23:27:02 crc kubenswrapper[4761]: I1125 23:27:02.752624 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-6z88v_09f2d102-2c78-4217-890e-39872745a907/registry-server/0.log" Nov 25 23:27:10 crc kubenswrapper[4761]: I1125 23:27:10.012477 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:27:10 crc kubenswrapper[4761]: E1125 23:27:10.013296 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:27:19 crc kubenswrapper[4761]: I1125 23:27:19.164478 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-t87pc_5f54a75a-fa05-4a82-b40f-eba6b9cc3af7/control-plane-machine-set-operator/0.log" Nov 25 23:27:19 crc kubenswrapper[4761]: I1125 23:27:19.373283 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pss5v_2aa3535f-06c8-4e56-ab10-589f653682aa/machine-api-operator/0.log" Nov 25 23:27:19 crc kubenswrapper[4761]: I1125 23:27:19.374957 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pss5v_2aa3535f-06c8-4e56-ab10-589f653682aa/kube-rbac-proxy/0.log" Nov 25 23:27:21 crc kubenswrapper[4761]: I1125 23:27:21.011195 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:27:21 crc kubenswrapper[4761]: E1125 23:27:21.012040 4761 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4lp9m_openshift-machine-config-operator(7e27d0af-be2c-4fc2-93f7-2f5294537b9d)\"" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" Nov 25 23:27:32 crc kubenswrapper[4761]: I1125 23:27:32.011906 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:27:32 crc kubenswrapper[4761]: I1125 23:27:32.831610 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"181a74cd55efe00de7061580856dc6ed53ad26f66f6d388f99845b88f20746e0"} Nov 25 23:27:36 crc kubenswrapper[4761]: I1125 23:27:36.981774 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-ftvpd_a6d03a2c-b764-4a2d-a55d-91cb3a608412/kube-rbac-proxy/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.078515 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-ftvpd_a6d03a2c-b764-4a2d-a55d-91cb3a608412/controller/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.148861 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-frr-files/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.256933 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-frr-files/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.264046 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-reloader/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.301340 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-reloader/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.359410 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-metrics/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.512744 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-frr-files/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.512695 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-reloader/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.553223 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-metrics/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.578121 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-metrics/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.701268 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-frr-files/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.709019 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-reloader/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.729475 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/cp-metrics/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.746233 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/controller/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.871995 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/frr-metrics/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.885184 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/kube-rbac-proxy/0.log" Nov 25 23:27:37 crc kubenswrapper[4761]: I1125 23:27:37.929131 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/kube-rbac-proxy-frr/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.120368 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/reloader/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.151037 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-shxv2_8220142c-79e5-43f5-9580-e88813af96cd/frr-k8s-webhook-server/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.320186 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d87477c79-vr8fq_bca3f822-b2ce-401e-b219-52d29e761c49/manager/3.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.363278 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d87477c79-vr8fq_bca3f822-b2ce-401e-b219-52d29e761c49/manager/2.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.391446 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l78lp_8973d4eb-45ec-49e9-a487-3963ae1b2fb9/frr/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.520854 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-55745dcbf5-4xft7_98ea4dd9-aefd-42d7-9a3a-c21aa5dbef23/webhook-server/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.542714 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sbf9j_a441d0d3-7366-499e-8096-2877837d1e58/kube-rbac-proxy/0.log" Nov 25 23:27:38 crc kubenswrapper[4761]: I1125 23:27:38.798231 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sbf9j_a441d0d3-7366-499e-8096-2877837d1e58/speaker/0.log" Nov 25 23:27:48 crc kubenswrapper[4761]: I1125 23:27:48.058614 4761 scope.go:117] "RemoveContainer" containerID="390337cc4cae0d91372d078a75e6fe818bf939c5f51c823ec46ae84d2e1618e7" Nov 25 23:27:48 crc kubenswrapper[4761]: I1125 23:27:48.095013 4761 scope.go:117] "RemoveContainer" containerID="079d882b2853b95d26ff0c45ae5e6a71e60a6d34a212f6a5fe1684be942deaf4" Nov 25 23:27:48 crc kubenswrapper[4761]: I1125 23:27:48.111018 4761 scope.go:117] "RemoveContainer" containerID="a9f8181244335bf706b8241ac2d5b4781c9d981d199747dcccfbcd1af841b226" Nov 25 23:27:53 crc kubenswrapper[4761]: I1125 23:27:53.761776 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-cdfb754f5-ddx65_d8846a3a-456b-480c-b911-6e205ff677ab/keystone-api/0.log" Nov 25 23:27:53 crc kubenswrapper[4761]: I1125 23:27:53.889876 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5cf936d6-7508-40de-a1ed-537da8087430/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.065564 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5cf936d6-7508-40de-a1ed-537da8087430/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.067667 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5cf936d6-7508-40de-a1ed-537da8087430/galera/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.254302 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_3c5a3c85-d95b-4252-8feb-155b5b742101/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.496856 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_3c5a3c85-d95b-4252-8feb-155b5b742101/galera/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.506729 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_3c5a3c85-d95b-4252-8feb-155b5b742101/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.681049 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d770ea4c-d11c-4526-af47-7507131606c0/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.819791 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d770ea4c-d11c-4526-af47-7507131606c0/mysql-bootstrap/0.log" Nov 25 23:27:54 crc kubenswrapper[4761]: I1125 23:27:54.878594 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d770ea4c-d11c-4526-af47-7507131606c0/galera/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.009287 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_77988aba-9a8b-4f63-b35a-6ecb7a4c8309/openstackclient/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.118859 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4370dc6c-2ed4-425a-bc57-98c19c37fde9/setup-container/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.307948 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4370dc6c-2ed4-425a-bc57-98c19c37fde9/setup-container/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.317327 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_a358e604-9896-4108-952a-6ad6560401ab/memcached/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.337582 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4370dc6c-2ed4-425a-bc57-98c19c37fde9/rabbitmq/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.448184 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-dxxvl_c7e65981-a172-41eb-ac4d-8bbe43248f0a/proxy-server/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.480660 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-dxxvl_c7e65981-a172-41eb-ac4d-8bbe43248f0a/proxy-httpd/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.517998 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-9mc86_b067022b-e70b-4750-9d89-f9e48ca02b04/swift-ring-rebalance/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.644686 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/account-auditor/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.669280 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/account-reaper/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.691221 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/account-replicator/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.839977 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/container-auditor/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.854522 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/account-server/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.856009 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/container-replicator/0.log" Nov 25 23:27:55 crc kubenswrapper[4761]: I1125 23:27:55.870448 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/container-server/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.008948 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/container-updater/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.048490 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/object-expirer/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.079418 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/object-auditor/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.081751 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/object-replicator/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.189338 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/object-server/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.260064 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/swift-recon-cron/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.284208 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/rsync/0.log" Nov 25 23:27:56 crc kubenswrapper[4761]: I1125 23:27:56.285712 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_ab080e3d-6c9d-4dd3-9b51-0b2ce3a86bad/object-updater/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.621709 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-utilities/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.749329 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-utilities/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.770981 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-content/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.796424 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-content/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.953219 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-utilities/0.log" Nov 25 23:28:10 crc kubenswrapper[4761]: I1125 23:28:10.955052 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/extract-content/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.207559 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-utilities/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.399392 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-content/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.399436 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-utilities/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.443896 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-content/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.452457 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fnwm8_534dc1b1-03d3-4b67-bbdc-a4ca72556281/registry-server/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.559333 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-utilities/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.620554 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/extract-content/0.log" Nov 25 23:28:11 crc kubenswrapper[4761]: I1125 23:28:11.799890 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/util/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.058806 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vcncj_9aacd34f-6d5f-4a1a-8a47-242df1d020d7/registry-server/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.160825 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/util/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.172807 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/pull/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.177412 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/pull/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.321781 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/util/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.369228 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/extract/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.373457 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6rnhx2_856fec2b-b44c-48c6-80a1-f71e87505db3/pull/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.524111 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-j2hzp_ff22925e-febe-45b5-8cf7-7de17a24c7f1/marketplace-operator/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.529198 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-utilities/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.713865 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-utilities/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.719447 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-content/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.765208 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-content/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.933341 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-content/0.log" Nov 25 23:28:12 crc kubenswrapper[4761]: I1125 23:28:12.949064 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/extract-utilities/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.083998 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-n9z85_23e39177-070d-4248-a455-7b2c98fd9486/registry-server/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.142450 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-utilities/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.289324 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-utilities/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.313719 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-content/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.318500 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-content/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.467190 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-content/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.467680 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/extract-utilities/0.log" Nov 25 23:28:13 crc kubenswrapper[4761]: I1125 23:28:13.859526 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h9c26_9e4155b4-dab7-4344-a1f7-cb95429b20ca/registry-server/0.log" Nov 25 23:29:20 crc kubenswrapper[4761]: I1125 23:29:20.801080 4761 generic.go:334] "Generic (PLEG): container finished" podID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerID="18667b8824eca762b66da1559cc361bee492b8665ff5de80dcdb7c86fd8204c9" exitCode=0 Nov 25 23:29:20 crc kubenswrapper[4761]: I1125 23:29:20.801209 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-99l4c/must-gather-jkfkv" event={"ID":"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5","Type":"ContainerDied","Data":"18667b8824eca762b66da1559cc361bee492b8665ff5de80dcdb7c86fd8204c9"} Nov 25 23:29:20 crc kubenswrapper[4761]: I1125 23:29:20.802633 4761 scope.go:117] "RemoveContainer" containerID="18667b8824eca762b66da1559cc361bee492b8665ff5de80dcdb7c86fd8204c9" Nov 25 23:29:20 crc kubenswrapper[4761]: I1125 23:29:20.920406 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-99l4c_must-gather-jkfkv_0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5/gather/0.log" Nov 25 23:29:27 crc kubenswrapper[4761]: I1125 23:29:27.670828 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-99l4c/must-gather-jkfkv"] Nov 25 23:29:27 crc kubenswrapper[4761]: I1125 23:29:27.671916 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-99l4c/must-gather-jkfkv" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="copy" containerID="cri-o://cf2fb587f81dca73cdc6ceca761d30460da5c708ca2c8cf2a19e3a229813efce" gracePeriod=2 Nov 25 23:29:27 crc kubenswrapper[4761]: I1125 23:29:27.676213 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-99l4c/must-gather-jkfkv"] Nov 25 23:29:27 crc kubenswrapper[4761]: I1125 23:29:27.868133 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-99l4c_must-gather-jkfkv_0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5/copy/0.log" Nov 25 23:29:27 crc kubenswrapper[4761]: I1125 23:29:27.868808 4761 generic.go:334] "Generic (PLEG): container finished" podID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerID="cf2fb587f81dca73cdc6ceca761d30460da5c708ca2c8cf2a19e3a229813efce" exitCode=143 Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.094586 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-99l4c_must-gather-jkfkv_0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5/copy/0.log" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.095018 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.227468 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output\") pod \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.227899 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgftx\" (UniqueName: \"kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx\") pod \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\" (UID: \"0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5\") " Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.234386 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx" (OuterVolumeSpecName: "kube-api-access-qgftx") pod "0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" (UID: "0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5"). InnerVolumeSpecName "kube-api-access-qgftx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.298276 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" (UID: "0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.330398 4761 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.330445 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgftx\" (UniqueName: \"kubernetes.io/projected/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5-kube-api-access-qgftx\") on node \"crc\" DevicePath \"\"" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.878606 4761 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-99l4c_must-gather-jkfkv_0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5/copy/0.log" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.879588 4761 scope.go:117] "RemoveContainer" containerID="cf2fb587f81dca73cdc6ceca761d30460da5c708ca2c8cf2a19e3a229813efce" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.879665 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-99l4c/must-gather-jkfkv" Nov 25 23:29:28 crc kubenswrapper[4761]: I1125 23:29:28.904241 4761 scope.go:117] "RemoveContainer" containerID="18667b8824eca762b66da1559cc361bee492b8665ff5de80dcdb7c86fd8204c9" Nov 25 23:29:29 crc kubenswrapper[4761]: I1125 23:29:29.025799 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" path="/var/lib/kubelet/pods/0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5/volumes" Nov 25 23:29:51 crc kubenswrapper[4761]: I1125 23:29:51.127938 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:29:51 crc kubenswrapper[4761]: I1125 23:29:51.128549 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.203085 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2"] Nov 25 23:30:00 crc kubenswrapper[4761]: E1125 23:30:00.204566 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="copy" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.204594 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="copy" Nov 25 23:30:00 crc kubenswrapper[4761]: E1125 23:30:00.204626 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="gather" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.204638 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="gather" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.204914 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="gather" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.204953 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4e86fd-dcd2-4cae-ad1f-16f4c7232ae5" containerName="copy" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.205744 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.209320 4761 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.211307 4761 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.231478 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2"] Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.255181 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.255350 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.255568 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2hnb\" (UniqueName: \"kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.357518 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2hnb\" (UniqueName: \"kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.357611 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.357671 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.359189 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.373886 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.383812 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2hnb\" (UniqueName: \"kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb\") pod \"collect-profiles-29401890-7srp2\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.539910 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:00 crc kubenswrapper[4761]: I1125 23:30:00.805826 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2"] Nov 25 23:30:01 crc kubenswrapper[4761]: I1125 23:30:01.248541 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" event={"ID":"00498099-3c3c-435f-b69f-fd58af846c53","Type":"ContainerStarted","Data":"0d6b6240b79204f620f38a7fc2e99bcd60ab29dd09e6c7422b565fa47b969e7f"} Nov 25 23:30:01 crc kubenswrapper[4761]: I1125 23:30:01.249295 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" event={"ID":"00498099-3c3c-435f-b69f-fd58af846c53","Type":"ContainerStarted","Data":"a75d458e7871c0d19bea49535f8a53dbc62ffccb119efd1135602d095bb73be0"} Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.259831 4761 generic.go:334] "Generic (PLEG): container finished" podID="00498099-3c3c-435f-b69f-fd58af846c53" containerID="0d6b6240b79204f620f38a7fc2e99bcd60ab29dd09e6c7422b565fa47b969e7f" exitCode=0 Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.259894 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" event={"ID":"00498099-3c3c-435f-b69f-fd58af846c53","Type":"ContainerDied","Data":"0d6b6240b79204f620f38a7fc2e99bcd60ab29dd09e6c7422b565fa47b969e7f"} Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.648056 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.795142 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume\") pod \"00498099-3c3c-435f-b69f-fd58af846c53\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.795640 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2hnb\" (UniqueName: \"kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb\") pod \"00498099-3c3c-435f-b69f-fd58af846c53\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.795727 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume\") pod \"00498099-3c3c-435f-b69f-fd58af846c53\" (UID: \"00498099-3c3c-435f-b69f-fd58af846c53\") " Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.796414 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume" (OuterVolumeSpecName: "config-volume") pod "00498099-3c3c-435f-b69f-fd58af846c53" (UID: "00498099-3c3c-435f-b69f-fd58af846c53"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.801953 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb" (OuterVolumeSpecName: "kube-api-access-q2hnb") pod "00498099-3c3c-435f-b69f-fd58af846c53" (UID: "00498099-3c3c-435f-b69f-fd58af846c53"). InnerVolumeSpecName "kube-api-access-q2hnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.805871 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "00498099-3c3c-435f-b69f-fd58af846c53" (UID: "00498099-3c3c-435f-b69f-fd58af846c53"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.898012 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2hnb\" (UniqueName: \"kubernetes.io/projected/00498099-3c3c-435f-b69f-fd58af846c53-kube-api-access-q2hnb\") on node \"crc\" DevicePath \"\"" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.898057 4761 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00498099-3c3c-435f-b69f-fd58af846c53-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:30:02 crc kubenswrapper[4761]: I1125 23:30:02.898075 4761 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00498099-3c3c-435f-b69f-fd58af846c53-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 23:30:03 crc kubenswrapper[4761]: I1125 23:30:03.270753 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" event={"ID":"00498099-3c3c-435f-b69f-fd58af846c53","Type":"ContainerDied","Data":"a75d458e7871c0d19bea49535f8a53dbc62ffccb119efd1135602d095bb73be0"} Nov 25 23:30:03 crc kubenswrapper[4761]: I1125 23:30:03.270804 4761 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a75d458e7871c0d19bea49535f8a53dbc62ffccb119efd1135602d095bb73be0" Nov 25 23:30:03 crc kubenswrapper[4761]: I1125 23:30:03.270915 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401890-7srp2" Nov 25 23:30:03 crc kubenswrapper[4761]: I1125 23:30:03.755360 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h"] Nov 25 23:30:03 crc kubenswrapper[4761]: I1125 23:30:03.767146 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401845-vbw4h"] Nov 25 23:30:05 crc kubenswrapper[4761]: I1125 23:30:05.027172 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c27c67d-99bf-4eea-bf3d-96a034d1e1b3" path="/var/lib/kubelet/pods/6c27c67d-99bf-4eea-bf3d-96a034d1e1b3/volumes" Nov 25 23:30:21 crc kubenswrapper[4761]: I1125 23:30:21.128317 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:30:21 crc kubenswrapper[4761]: I1125 23:30:21.129006 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:30:48 crc kubenswrapper[4761]: I1125 23:30:48.223570 4761 scope.go:117] "RemoveContainer" containerID="e226dedb2b473e1e6997ad3f1d676f4caba84e619dc5d4da7886acd0e7d51a13" Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.128567 4761 patch_prober.go:28] interesting pod/machine-config-daemon-4lp9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.129992 4761 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.130332 4761 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.131130 4761 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"181a74cd55efe00de7061580856dc6ed53ad26f66f6d388f99845b88f20746e0"} pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.131452 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" podUID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerName="machine-config-daemon" containerID="cri-o://181a74cd55efe00de7061580856dc6ed53ad26f66f6d388f99845b88f20746e0" gracePeriod=600 Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.735493 4761 generic.go:334] "Generic (PLEG): container finished" podID="7e27d0af-be2c-4fc2-93f7-2f5294537b9d" containerID="181a74cd55efe00de7061580856dc6ed53ad26f66f6d388f99845b88f20746e0" exitCode=0 Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.735602 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerDied","Data":"181a74cd55efe00de7061580856dc6ed53ad26f66f6d388f99845b88f20746e0"} Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.736122 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4lp9m" event={"ID":"7e27d0af-be2c-4fc2-93f7-2f5294537b9d","Type":"ContainerStarted","Data":"c28d0b53f5efb7fb27d88dca0f0fe611e04d34c5372cfe99a7059e83a814f483"} Nov 25 23:30:51 crc kubenswrapper[4761]: I1125 23:30:51.736143 4761 scope.go:117] "RemoveContainer" containerID="cce560213a0c8e0f5e5220d9c4fad5f56cd2728e2b01512eef4c527ae3d0b251" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.840272 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:37 crc kubenswrapper[4761]: E1125 23:31:37.841572 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00498099-3c3c-435f-b69f-fd58af846c53" containerName="collect-profiles" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.841597 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="00498099-3c3c-435f-b69f-fd58af846c53" containerName="collect-profiles" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.841836 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="00498099-3c3c-435f-b69f-fd58af846c53" containerName="collect-profiles" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.846854 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.850469 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw7bf\" (UniqueName: \"kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.850528 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.850579 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.897291 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.951993 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.952357 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw7bf\" (UniqueName: \"kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.952434 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.952834 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.952852 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:37 crc kubenswrapper[4761]: I1125 23:31:37.975242 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw7bf\" (UniqueName: \"kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf\") pod \"certified-operators-8645l\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:38 crc kubenswrapper[4761]: I1125 23:31:38.216616 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:38 crc kubenswrapper[4761]: I1125 23:31:38.669758 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:38 crc kubenswrapper[4761]: W1125 23:31:38.680564 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d7c1b2a_dec5_44b7_9d8e_cc3feb208ac8.slice/crio-79439385959d447d6938367a9bc22dbc3b2beea8d14c4fcc81e136b2a0e41bb5 WatchSource:0}: Error finding container 79439385959d447d6938367a9bc22dbc3b2beea8d14c4fcc81e136b2a0e41bb5: Status 404 returned error can't find the container with id 79439385959d447d6938367a9bc22dbc3b2beea8d14c4fcc81e136b2a0e41bb5 Nov 25 23:31:39 crc kubenswrapper[4761]: I1125 23:31:39.425044 4761 generic.go:334] "Generic (PLEG): container finished" podID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerID="fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f" exitCode=0 Nov 25 23:31:39 crc kubenswrapper[4761]: I1125 23:31:39.425173 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerDied","Data":"fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f"} Nov 25 23:31:39 crc kubenswrapper[4761]: I1125 23:31:39.425405 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerStarted","Data":"79439385959d447d6938367a9bc22dbc3b2beea8d14c4fcc81e136b2a0e41bb5"} Nov 25 23:31:39 crc kubenswrapper[4761]: I1125 23:31:39.428000 4761 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.600673 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.611453 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.619718 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.799577 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhwb\" (UniqueName: \"kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.799634 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.799670 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.901464 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhwb\" (UniqueName: \"kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.901531 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.901569 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.902146 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.902376 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.927132 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhwb\" (UniqueName: \"kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb\") pod \"community-operators-wp27c\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:40 crc kubenswrapper[4761]: I1125 23:31:40.930767 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:41 crc kubenswrapper[4761]: I1125 23:31:41.411878 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:41 crc kubenswrapper[4761]: W1125 23:31:41.417855 4761 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51e555ac_6e64_4371_803c_871ed48ff7c6.slice/crio-04d88ecfc247170f87134e6b5383fa71ef4618e5b5fb4de32be35230ba8a1c7a WatchSource:0}: Error finding container 04d88ecfc247170f87134e6b5383fa71ef4618e5b5fb4de32be35230ba8a1c7a: Status 404 returned error can't find the container with id 04d88ecfc247170f87134e6b5383fa71ef4618e5b5fb4de32be35230ba8a1c7a Nov 25 23:31:41 crc kubenswrapper[4761]: I1125 23:31:41.445770 4761 generic.go:334] "Generic (PLEG): container finished" podID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerID="312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b" exitCode=0 Nov 25 23:31:41 crc kubenswrapper[4761]: I1125 23:31:41.445867 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerDied","Data":"312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b"} Nov 25 23:31:41 crc kubenswrapper[4761]: I1125 23:31:41.449246 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerStarted","Data":"04d88ecfc247170f87134e6b5383fa71ef4618e5b5fb4de32be35230ba8a1c7a"} Nov 25 23:31:42 crc kubenswrapper[4761]: I1125 23:31:42.461505 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerStarted","Data":"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a"} Nov 25 23:31:42 crc kubenswrapper[4761]: I1125 23:31:42.465878 4761 generic.go:334] "Generic (PLEG): container finished" podID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerID="5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c" exitCode=0 Nov 25 23:31:42 crc kubenswrapper[4761]: I1125 23:31:42.465940 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerDied","Data":"5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c"} Nov 25 23:31:42 crc kubenswrapper[4761]: I1125 23:31:42.497041 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8645l" podStartSLOduration=3.029047922 podStartE2EDuration="5.497016835s" podCreationTimestamp="2025-11-25 23:31:37 +0000 UTC" firstStartedPulling="2025-11-25 23:31:39.427615153 +0000 UTC m=+2095.160960998" lastFinishedPulling="2025-11-25 23:31:41.895584076 +0000 UTC m=+2097.628929911" observedRunningTime="2025-11-25 23:31:42.489908355 +0000 UTC m=+2098.223254230" watchObservedRunningTime="2025-11-25 23:31:42.497016835 +0000 UTC m=+2098.230362710" Nov 25 23:31:43 crc kubenswrapper[4761]: I1125 23:31:43.494904 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerStarted","Data":"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016"} Nov 25 23:31:44 crc kubenswrapper[4761]: I1125 23:31:44.505177 4761 generic.go:334] "Generic (PLEG): container finished" podID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerID="4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016" exitCode=0 Nov 25 23:31:44 crc kubenswrapper[4761]: I1125 23:31:44.505247 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerDied","Data":"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016"} Nov 25 23:31:45 crc kubenswrapper[4761]: I1125 23:31:45.515394 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerStarted","Data":"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb"} Nov 25 23:31:45 crc kubenswrapper[4761]: I1125 23:31:45.544169 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wp27c" podStartSLOduration=3.096743661 podStartE2EDuration="5.544145184s" podCreationTimestamp="2025-11-25 23:31:40 +0000 UTC" firstStartedPulling="2025-11-25 23:31:42.469239854 +0000 UTC m=+2098.202585719" lastFinishedPulling="2025-11-25 23:31:44.916641407 +0000 UTC m=+2100.649987242" observedRunningTime="2025-11-25 23:31:45.543803644 +0000 UTC m=+2101.277149569" watchObservedRunningTime="2025-11-25 23:31:45.544145184 +0000 UTC m=+2101.277491049" Nov 25 23:31:48 crc kubenswrapper[4761]: I1125 23:31:48.217651 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:48 crc kubenswrapper[4761]: I1125 23:31:48.217970 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:48 crc kubenswrapper[4761]: I1125 23:31:48.282637 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:48 crc kubenswrapper[4761]: I1125 23:31:48.294735 4761 scope.go:117] "RemoveContainer" containerID="6f36081412228b01edd95f9e1fd24113d067f09f15a3335e5b91037a70fa9254" Nov 25 23:31:48 crc kubenswrapper[4761]: I1125 23:31:48.601496 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:49 crc kubenswrapper[4761]: I1125 23:31:49.388384 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:50 crc kubenswrapper[4761]: I1125 23:31:50.557260 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8645l" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="registry-server" containerID="cri-o://a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a" gracePeriod=2 Nov 25 23:31:50 crc kubenswrapper[4761]: I1125 23:31:50.931482 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:50 crc kubenswrapper[4761]: I1125 23:31:50.931540 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.006213 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.034984 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.055095 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities\") pod \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.055319 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw7bf\" (UniqueName: \"kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf\") pod \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.055445 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content\") pod \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\" (UID: \"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8\") " Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.055826 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities" (OuterVolumeSpecName: "utilities") pod "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" (UID: "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.056248 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.066229 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf" (OuterVolumeSpecName: "kube-api-access-hw7bf") pod "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" (UID: "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8"). InnerVolumeSpecName "kube-api-access-hw7bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.122202 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" (UID: "9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.158204 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.158228 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw7bf\" (UniqueName: \"kubernetes.io/projected/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8-kube-api-access-hw7bf\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.567684 4761 generic.go:334] "Generic (PLEG): container finished" podID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerID="a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a" exitCode=0 Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.567863 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8645l" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.567852 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerDied","Data":"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a"} Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.568078 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8645l" event={"ID":"9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8","Type":"ContainerDied","Data":"79439385959d447d6938367a9bc22dbc3b2beea8d14c4fcc81e136b2a0e41bb5"} Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.568123 4761 scope.go:117] "RemoveContainer" containerID="a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.600588 4761 scope.go:117] "RemoveContainer" containerID="312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.624958 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.630551 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8645l"] Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.640295 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.648869 4761 scope.go:117] "RemoveContainer" containerID="fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.669671 4761 scope.go:117] "RemoveContainer" containerID="a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a" Nov 25 23:31:51 crc kubenswrapper[4761]: E1125 23:31:51.670358 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a\": container with ID starting with a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a not found: ID does not exist" containerID="a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.670452 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a"} err="failed to get container status \"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a\": rpc error: code = NotFound desc = could not find container \"a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a\": container with ID starting with a15e4f6dd25000289edcbab714c9ce6b48deeb1759cfbcece0fd7ca77426db7a not found: ID does not exist" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.670498 4761 scope.go:117] "RemoveContainer" containerID="312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b" Nov 25 23:31:51 crc kubenswrapper[4761]: E1125 23:31:51.670855 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b\": container with ID starting with 312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b not found: ID does not exist" containerID="312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.670906 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b"} err="failed to get container status \"312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b\": rpc error: code = NotFound desc = could not find container \"312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b\": container with ID starting with 312474b3ae0eb02a8332d8df9cd5ea581c5d5151c0fb425bff72f425df08b31b not found: ID does not exist" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.670931 4761 scope.go:117] "RemoveContainer" containerID="fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f" Nov 25 23:31:51 crc kubenswrapper[4761]: E1125 23:31:51.671426 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f\": container with ID starting with fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f not found: ID does not exist" containerID="fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f" Nov 25 23:31:51 crc kubenswrapper[4761]: I1125 23:31:51.671468 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f"} err="failed to get container status \"fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f\": rpc error: code = NotFound desc = could not find container \"fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f\": container with ID starting with fc549351eb3b4a3418eff98041f70679b1a68726142353cf2c779b32d595571f not found: ID does not exist" Nov 25 23:31:53 crc kubenswrapper[4761]: I1125 23:31:53.027607 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" path="/var/lib/kubelet/pods/9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8/volumes" Nov 25 23:31:53 crc kubenswrapper[4761]: I1125 23:31:53.390047 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:53 crc kubenswrapper[4761]: I1125 23:31:53.590894 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wp27c" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="registry-server" containerID="cri-o://5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb" gracePeriod=2 Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.159884 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.229166 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwhwb\" (UniqueName: \"kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb\") pod \"51e555ac-6e64-4371-803c-871ed48ff7c6\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.229277 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content\") pod \"51e555ac-6e64-4371-803c-871ed48ff7c6\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.229324 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities\") pod \"51e555ac-6e64-4371-803c-871ed48ff7c6\" (UID: \"51e555ac-6e64-4371-803c-871ed48ff7c6\") " Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.230523 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities" (OuterVolumeSpecName: "utilities") pod "51e555ac-6e64-4371-803c-871ed48ff7c6" (UID: "51e555ac-6e64-4371-803c-871ed48ff7c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.234632 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb" (OuterVolumeSpecName: "kube-api-access-fwhwb") pod "51e555ac-6e64-4371-803c-871ed48ff7c6" (UID: "51e555ac-6e64-4371-803c-871ed48ff7c6"). InnerVolumeSpecName "kube-api-access-fwhwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.285976 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51e555ac-6e64-4371-803c-871ed48ff7c6" (UID: "51e555ac-6e64-4371-803c-871ed48ff7c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.330672 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.330733 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwhwb\" (UniqueName: \"kubernetes.io/projected/51e555ac-6e64-4371-803c-871ed48ff7c6-kube-api-access-fwhwb\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.330746 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e555ac-6e64-4371-803c-871ed48ff7c6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.604968 4761 generic.go:334] "Generic (PLEG): container finished" podID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerID="5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb" exitCode=0 Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.605029 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerDied","Data":"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb"} Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.605035 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp27c" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.605114 4761 scope.go:117] "RemoveContainer" containerID="5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.605076 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp27c" event={"ID":"51e555ac-6e64-4371-803c-871ed48ff7c6","Type":"ContainerDied","Data":"04d88ecfc247170f87134e6b5383fa71ef4618e5b5fb4de32be35230ba8a1c7a"} Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.636970 4761 scope.go:117] "RemoveContainer" containerID="4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.667317 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.674807 4761 scope.go:117] "RemoveContainer" containerID="5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.677996 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wp27c"] Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.716300 4761 scope.go:117] "RemoveContainer" containerID="5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb" Nov 25 23:31:54 crc kubenswrapper[4761]: E1125 23:31:54.716940 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb\": container with ID starting with 5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb not found: ID does not exist" containerID="5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.716996 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb"} err="failed to get container status \"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb\": rpc error: code = NotFound desc = could not find container \"5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb\": container with ID starting with 5129d858dd403ee9eaa6386d4a3f3f21574579a808a510a7eafd056a580ffdeb not found: ID does not exist" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.717030 4761 scope.go:117] "RemoveContainer" containerID="4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016" Nov 25 23:31:54 crc kubenswrapper[4761]: E1125 23:31:54.717470 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016\": container with ID starting with 4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016 not found: ID does not exist" containerID="4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.717510 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016"} err="failed to get container status \"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016\": rpc error: code = NotFound desc = could not find container \"4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016\": container with ID starting with 4dbb3b919cd68e3daa4228da2889371b5da3dd21b7cc23b608bf6d70f95cc016 not found: ID does not exist" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.717544 4761 scope.go:117] "RemoveContainer" containerID="5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c" Nov 25 23:31:54 crc kubenswrapper[4761]: E1125 23:31:54.717958 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c\": container with ID starting with 5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c not found: ID does not exist" containerID="5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c" Nov 25 23:31:54 crc kubenswrapper[4761]: I1125 23:31:54.717999 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c"} err="failed to get container status \"5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c\": rpc error: code = NotFound desc = could not find container \"5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c\": container with ID starting with 5894ec8cd0692c866fb7431627c1f607e598f45b050d943ee179aeaad5af9f8c not found: ID does not exist" Nov 25 23:31:55 crc kubenswrapper[4761]: I1125 23:31:55.026542 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" path="/var/lib/kubelet/pods/51e555ac-6e64-4371-803c-871ed48ff7c6/volumes" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.050278 4761 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051155 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="extract-utilities" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051171 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="extract-utilities" Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051189 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="extract-content" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051197 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="extract-content" Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051218 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="extract-content" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051227 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="extract-content" Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051244 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051252 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051267 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051276 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: E1125 23:31:59.051293 4761 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="extract-utilities" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051301 4761 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="extract-utilities" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051462 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e555ac-6e64-4371-803c-871ed48ff7c6" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.051479 4761 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d7c1b2a-dec5-44b7-9d8e-cc3feb208ac8" containerName="registry-server" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.052782 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.069262 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.212502 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wlhh\" (UniqueName: \"kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.212599 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.212658 4761 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.314176 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.314319 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wlhh\" (UniqueName: \"kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.314353 4761 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.314878 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.315134 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.338762 4761 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wlhh\" (UniqueName: \"kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh\") pod \"redhat-operators-b8r5g\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.371981 4761 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:31:59 crc kubenswrapper[4761]: I1125 23:31:59.807470 4761 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:32:00 crc kubenswrapper[4761]: I1125 23:32:00.658629 4761 generic.go:334] "Generic (PLEG): container finished" podID="4c0be87b-668c-42f5-8c72-c00ee7ccec50" containerID="0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c" exitCode=0 Nov 25 23:32:00 crc kubenswrapper[4761]: I1125 23:32:00.658985 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerDied","Data":"0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c"} Nov 25 23:32:00 crc kubenswrapper[4761]: I1125 23:32:00.659024 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerStarted","Data":"5c0919bd15716f9b1611b089a5a39da91d711c5cc01a4e9630191b5fea9af683"} Nov 25 23:32:01 crc kubenswrapper[4761]: I1125 23:32:01.667802 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerStarted","Data":"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850"} Nov 25 23:32:02 crc kubenswrapper[4761]: I1125 23:32:02.679147 4761 generic.go:334] "Generic (PLEG): container finished" podID="4c0be87b-668c-42f5-8c72-c00ee7ccec50" containerID="4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850" exitCode=0 Nov 25 23:32:02 crc kubenswrapper[4761]: I1125 23:32:02.679236 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerDied","Data":"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850"} Nov 25 23:32:03 crc kubenswrapper[4761]: I1125 23:32:03.691672 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerStarted","Data":"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478"} Nov 25 23:32:03 crc kubenswrapper[4761]: I1125 23:32:03.720223 4761 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b8r5g" podStartSLOduration=2.312000321 podStartE2EDuration="4.720190087s" podCreationTimestamp="2025-11-25 23:31:59 +0000 UTC" firstStartedPulling="2025-11-25 23:32:00.665847847 +0000 UTC m=+2116.399193682" lastFinishedPulling="2025-11-25 23:32:03.074037603 +0000 UTC m=+2118.807383448" observedRunningTime="2025-11-25 23:32:03.716744735 +0000 UTC m=+2119.450090650" watchObservedRunningTime="2025-11-25 23:32:03.720190087 +0000 UTC m=+2119.453535992" Nov 25 23:32:09 crc kubenswrapper[4761]: I1125 23:32:09.372313 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:09 crc kubenswrapper[4761]: I1125 23:32:09.374271 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:09 crc kubenswrapper[4761]: I1125 23:32:09.428590 4761 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:09 crc kubenswrapper[4761]: I1125 23:32:09.807825 4761 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:09 crc kubenswrapper[4761]: I1125 23:32:09.854654 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:32:11 crc kubenswrapper[4761]: I1125 23:32:11.764524 4761 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b8r5g" podUID="4c0be87b-668c-42f5-8c72-c00ee7ccec50" containerName="registry-server" containerID="cri-o://5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478" gracePeriod=2 Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.547475 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.655552 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content\") pod \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.655652 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities\") pod \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.655844 4761 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wlhh\" (UniqueName: \"kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh\") pod \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\" (UID: \"4c0be87b-668c-42f5-8c72-c00ee7ccec50\") " Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.657051 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities" (OuterVolumeSpecName: "utilities") pod "4c0be87b-668c-42f5-8c72-c00ee7ccec50" (UID: "4c0be87b-668c-42f5-8c72-c00ee7ccec50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.679307 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh" (OuterVolumeSpecName: "kube-api-access-9wlhh") pod "4c0be87b-668c-42f5-8c72-c00ee7ccec50" (UID: "4c0be87b-668c-42f5-8c72-c00ee7ccec50"). InnerVolumeSpecName "kube-api-access-9wlhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.756465 4761 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c0be87b-668c-42f5-8c72-c00ee7ccec50" (UID: "4c0be87b-668c-42f5-8c72-c00ee7ccec50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.757357 4761 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.757483 4761 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0be87b-668c-42f5-8c72-c00ee7ccec50-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.757542 4761 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wlhh\" (UniqueName: \"kubernetes.io/projected/4c0be87b-668c-42f5-8c72-c00ee7ccec50-kube-api-access-9wlhh\") on node \"crc\" DevicePath \"\"" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.804212 4761 generic.go:334] "Generic (PLEG): container finished" podID="4c0be87b-668c-42f5-8c72-c00ee7ccec50" containerID="5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478" exitCode=0 Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.804274 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerDied","Data":"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478"} Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.804316 4761 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b8r5g" event={"ID":"4c0be87b-668c-42f5-8c72-c00ee7ccec50","Type":"ContainerDied","Data":"5c0919bd15716f9b1611b089a5a39da91d711c5cc01a4e9630191b5fea9af683"} Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.804338 4761 scope.go:117] "RemoveContainer" containerID="5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.804523 4761 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b8r5g" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.828274 4761 scope.go:117] "RemoveContainer" containerID="4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.841813 4761 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.850093 4761 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b8r5g"] Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.870974 4761 scope.go:117] "RemoveContainer" containerID="0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.897103 4761 scope.go:117] "RemoveContainer" containerID="5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478" Nov 25 23:32:13 crc kubenswrapper[4761]: E1125 23:32:13.897654 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478\": container with ID starting with 5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478 not found: ID does not exist" containerID="5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.897804 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478"} err="failed to get container status \"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478\": rpc error: code = NotFound desc = could not find container \"5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478\": container with ID starting with 5b3a3df003adc348c2facf34876043ae438e245c4ea9d3d48c4437ce0547c478 not found: ID does not exist" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.897833 4761 scope.go:117] "RemoveContainer" containerID="4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850" Nov 25 23:32:13 crc kubenswrapper[4761]: E1125 23:32:13.899575 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850\": container with ID starting with 4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850 not found: ID does not exist" containerID="4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.899609 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850"} err="failed to get container status \"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850\": rpc error: code = NotFound desc = could not find container \"4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850\": container with ID starting with 4f8d3ae78bca623134cae8e1cb20593e7d53f11df505d8e908f8d11aeba6e850 not found: ID does not exist" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.899628 4761 scope.go:117] "RemoveContainer" containerID="0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c" Nov 25 23:32:13 crc kubenswrapper[4761]: E1125 23:32:13.899996 4761 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c\": container with ID starting with 0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c not found: ID does not exist" containerID="0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c" Nov 25 23:32:13 crc kubenswrapper[4761]: I1125 23:32:13.900026 4761 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c"} err="failed to get container status \"0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c\": rpc error: code = NotFound desc = could not find container \"0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c\": container with ID starting with 0bcaab08c315ba47fc34e82a028e3d8c06bec9c9b2fcc0ae1d9ff09ea10ef84c not found: ID does not exist" Nov 25 23:32:15 crc kubenswrapper[4761]: I1125 23:32:15.024139 4761 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c0be87b-668c-42f5-8c72-c00ee7ccec50" path="/var/lib/kubelet/pods/4c0be87b-668c-42f5-8c72-c00ee7ccec50/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111436231024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111436232017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111431576016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111431576015461 5ustar corecore